Uneingeschränkter Zugang

AIGC-enabled Education Information Technology Integration Application and Research--Taking Information Technology Teaching of Preschool Education Major as an Example

  
24. März 2025

Zitieren
COVER HERUNTERLADEN

Introduction

Currently, Artificial Intelligence Generated Content (AIGC) represented by Chat GPT has aroused the attention of all walks of life, which provides the possibility for users to create high-quality content quickly, and its technological advantages have promoted the integration of various fields with it [1-3]. Digital technology has promoted the advancement of teaching methods, teaching tools, and teaching techniques, and the informatization of education has had a positive impact on the learning effectiveness of students, the quality of teachers’ teaching, and the management efficiency of educational institutions [4-6]. However, the development of new technologies brings both opportunities and challenges, especially in the field of education informatization, and the arrival of AIGC heralds a change in the way of education [7-8].

The phenomenon of data silos in the process of education informatization is widespread, and this problem hinders the effective sharing and circulation of information, which in turn leads to the problem of data duplication and inconsistency [9]. AIGC, as a technology with the ability to integrate, can build an efficient teaching, research and management platform by integrating all kinds of educational resources and services, so as to enhance the optimization of the degree of resource allocation and the efficiency of sharing, and make a significant contribution to the contribute to promoting educational equity and improving educational quality [10-13]. In the application practice of education informatization, AIGC’s customization capability can be used to innovate the teaching mode, and teachers are able to personalize the design according to students’ learning needs [14-15]. At the research level, AIGC’s customization capability can intelligently push the latest research results and project funding information according to researchers’ research fields and interests [16-17]. At the level of education management, AIGC is able to provide administrators with customized data analysis and decision support, thus enabling more efficient resource allocation [18-19]. In order to maximize the potential of AIGC technology, it is important to comprehensively consider the challenges it poses and adopt effective strategies to deal with them, which requires the concerted efforts of educational institutions, policy makers, technology developers, and the community.

The introduction of AIGC technology provides strong support for the design and implementation of personalized learning experiences. In order to explore the integration application of AIGC-enabled education information technology, this paper firstly takes the teaching of information technology in preschool education as an example to sort out the application of AIGC technology in preschool education majors. Secondly, the method of generating mind maps based on teaching videos is proposed, and the generated mind maps are used as tools to design classroom teaching in preschool education majors. Considering the use of teaching resource knowledge graph to complete the recommendation task, DB-CGAT model is proposed to combine the teaching resource knowledge graph context processing method with the dual behavior aggregation method. Finally, the effectiveness of the teaching method based on mind maps is verified by the questionnaire method. The CoLR dataset for teaching resources has been constructed, and the validity verification of the DB-CGAT model has been completed in conjunction with the existing Yelp2018 and Amazon-Book datasets.

Application of AIGC technology in preschool education programs

In the field of preschool education, the innovative generation of teaching resources plays a decisive role in improving the quality of education and the effectiveness of learning, and the introduction of AIGC technology has not only led to a fundamental change in the way teaching resources are created, distributed, and consumed, but also provided strong support for the design and implementation of personalized learning experiences.

The Role of AIGC Technology in Teaching Resource Innovation

AIGC technology, through its automation mechanism, provides teachers with diverse and customized teaching resources to better meet students’ learning needs, and its ability to personalize learning content according to students’ learning progress and comprehension enhances learning efficiency and motivation.AIGC technology goes beyond the traditional resources that are limited to text and static images, and is able to create content such as videos, audios, and animations. This rich content not only adds interest to learning, but also promotes students’ understanding and memorization of complex concepts. In addition, AIGC technology can be used to develop Virtual Reality (VR) and Augmented Reality (AR) applications that offer students immersive learning experiences. In order to adapt to changing industry standards and skill needs in preschool education, AIGC technology also facilitates rapid iteration and updating of instructional resources, and its automated generation and updating mechanism improves the speed and flexibility of updating instructional resources. Teachers can use AIGC technology to quickly create instructional materials that reflect the latest industry trends and skill needs, ensuring that the content is current and relevant.

Development of personalized learning paths

In the context of contemporary preschool education, the development of personalized learning paths has become crucial, and AIGC technology precisely meets the personalized learning needs of students through in-depth analyses of their learning behaviors, cognitive styles, and achievement levels, and by tailoring learning content and activities for them. Personalized learning content generation is another important application of AIGC technology. Using machine learning and data mining algorithms, AIGC technology is able to analyze students’ learning data, identify learning styles and interests, and generate learning materials that match their profiles. This personalized content not only improves learning efficiency, but also significantly increases student motivation and engagement. For visual learners, AIGC technology generates learning materials with graphics, videos, and animations. For learners who prefer text, AIGC technology provides text content with detailed explanations and case studies. Another significant advantage of AIGC technology is its ability to dynamically adjust the learning path according to students’ progress and comprehension. By monitoring students’ performance in real time, AIGC technology is able to adjust the difficulty and complexity of the learning materials to ensure that students are always at the right level of challenge. This dynamic adjustment mechanism not only prevents students from becoming frustrated and bored when the material is too difficult or too easy, but also greatly facilitates the realization of optimal learning outcomes.AIGC technology is also used to create highly interactive learning experiences. By combining natural language processing and speech recognition technologies, AIGC technology is capable of interacting with students in fluent natural language, answering questions effectively, providing feedback, and even simulated real-world situations. This interactive learning experience not only enhances student engagement, but also helps develop their critical thinking and problem-solving skills.

AIGC technology-driven preschool teaching and learning
Methods for generating mind maps based on instructional videos

The mind map generation method is a method of using technical means to transform text data into a mind map, which reduces the time spent on reading text and the process of constructing a mind map compared with the manual or software way of drawing a mind map. This generation method gives full play to the characterization function of the mind map, which facilitates users to directly understand the main knowledge structure of the text content, thus deepening the memory and understanding of the learning content.

Methodological framework

In order to quickly understand what specific knowledge points are contained in the teaching video and help users construct a knowledge framework based on the teaching video, this study proposes a mind map generation method based on the teaching video. To achieve this goal, this study integrates the results of previous research and aims to realize the generation of mind maps based on teaching videos through multiple processing stages. In the first step, the teaching video is pre-prepared to get the text data under different topics; in the second step, the data is pre-processed with the aim of getting the content summaries under the texts of different topics; in the third step, in order to establish the relationship between the keywords, the sub-texts are subjected to the textual syntactic dependency analysis and semantic dependency analysis, so as to get the hierarchical structure of the keywords; finally, the obtained topics, sub-topics, and semantically linked keywords are visualized and generated to get the mind map illustration. The idea of generating mind maps using teaching videos is shown in Figure 1.

Figure 1.

Mind Map Generation Method Based on Teaching Video

Data pre-processing

For the text data obtained from transcription, the next step is mainly to summarize the transcribed text information and obtain text summaries. Data preprocessing is a necessary operation before analyzing Chinese text. It refers to obtaining more reliable data through Chinese word splitting and other operations, correcting errors and abnormal data, and improving the accuracy of analysis. The preprocessing of text will directly affect the quality of the subsequent text analysis. Therefore, an efficient preprocessing operation in Chinese text is an important condition to ensure that an accurate mind map is obtained. The specific process of data preprocessing is depicted in Figure 2.

Figure 2.

Data preprocessing process

Text summary extraction

Text summary extraction refers to the technique of converting text content into a short summary that contains key information. At present, extractive and generative methods are the two main types of automatic text summary extraction.

The main idea of extractive text summarization is to rely on the original text, with a number of sentences before and after the original text to indicate the summary, the specific approach is to split the text into sentences, and according to certain rules to calculate the weight of each sentence, the weight of a number of sentences with the highest ranking as a text summary. The generative text summarization algorithm, on the other hand, needs to understand and summarize the text content. With the deep development of deep learning technology, the Seq2Seq model has excellent performance in the generative text summary extraction task. However, its generation results are not stable, and due to the complexity and non-interpretability of the model, it cannot meet the requirements of practical applications at present. Therefore, extractive-based text summarization methods continue to be widely popular.

This study mainly focuses on quickly obtaining the summaries in the subtext data obtained in the previous section through the TF-IDF algorithm and the TextRank algorithm, with the aim of eliminating disposable or poorly related statements and analyzing the main content specifically covered in each subtopic. By comparing the TF-IDF algorithm and the TextRank algorithm, it is found that the former is more dependent on the corpus environment, and therefore has more statistical advantages; while the latter further considers the semantic relationship between words within the document, and has the advantage of reflecting the co-occurrence information between words within the document. Therefore, the TF-IDF algorithm and TextRank have their own advantages and disadvantages, and the two algorithms will be used comprehensively in this study to compare each other and improve the accuracy of text summary extraction.

TF-IDF algorithm

TF-IDF is an algorithm based on statistical analysis to indicate the importance of a word or sentence in a text. A simple idea is to find the word with the highest frequency of occurrence, and the more frequently it is used, the more important it is indicated. Thus, the concept of word frequency is introduced. However, the most frequent words in a text are likely to be “of”, ‘“in”, “is”, etc. This is a common category of meaningless words, so we need to filter out these words in the actual task. These are commonly used meaningless words, so we need to filter them out in the actual task. However, when only words with actual meaning are left, we find that the words “computer”, “hardware”, and “memory” appear as often as they do in a computer text. Obviously they are not of the same importance. This is because “computer” is a very common word in computer texts. Relatively speaking, “hardware” and “memory” are less common. This also means that “hardware” and “memory” are more important than “computer”, so in the keyword ranking, “hardware” and “memory” should be ranked before “computer”.

Therefore, the concept of reverse document frequency was introduced to measure how common a word is. More common words, such as “computer”, are given less weight, while less common words, such as “memory”, are given more weight.

TF-IDF value is the product of TF and IDF value, the larger the TF-IDF value, the more important the word, we simply call it “keyword”. The importance of a sentence can be measured by its “keywords”. If a sentence contains more “keywords”, the more important it is. The N sentences with the highest ranked importance level are selected to form a text summary. In conclusion, TF-IDF is a typical automatic summarization algorithm based on word frequency statistics. The specific formula is shown below.

Word frequency: the frequency of keyword w in the text Di. Where count(w) indicates the number of occurrences of keyword w, |Di|$$\left| {{D_i}} \right|$$ indicates the total number of words in the text. TFw,Di=count(w)|Di|$$T{F_{w,{D_i}}} = \frac{{count\left( w \right)}}{{\left| {{D_i}} \right|}}$$

Inverse Document Frequency: It indicates how common the word is. The denominator indicates the number of documents containing the keyword; the more common a word is, the larger the denominator and the smaller the IDF value. IDFw=logN1+i=1MI(w,Di)$$ID{F_w} = \log \frac{N}{{1 + \sum\limits_{i = 1}^M {I\left( {w,{D_i}} \right)} }}$$

By invoking the TF-IDF algorithm to extract text summaries for the text under each sub-topic separately, the main content under the topic can be obtained. For example, if we call the TF-IDF algorithm on the text under the topic of “Memory”, we can get the summary of the text as “So the second major component is the memory, which is also the part of the computer used to store the program and data. So there are two main parts of the memory, one is the internal memory and the other is the external memory.”

TextRank Algorithm

TextRank algorithm was proposed by Mihalcea and Tarau in 2004 as an algorithm that can be used as keyword extraction as well as phrase extraction and summary extraction. The algorithm is inspired by the PageRank algorithm, which assumes that the Internet is a directed graph and the nodes are web pages, and determines the importance of a web page based on the probability of linking to it and being linked to it. Similarly, the TextRank algorithm sees the document as a network of constituent units (words or sentences), and the semantic relationships between words (sentences) and words (sentences) are used as links in the network, and the ranking of the words is obtained by constructing a candidate keyword graph, and the highly ranked words are used as text keywords, and the key phrases as well as the text summaries are obtained based on the number of keywords.

TextRank algorithm extracts the text summary process: first, according to the need to integrate the contents of multiple articles into the same text, and split them into sentences, these sentences are represented as vectors; after that, the similarity between two sentences is calculated, and the similarity score is used as the weight of the edges between the two sentences, and the scores of each sentence are obtained by iteratively calculating the weights of the edges, and finally the sentences are ranked in accordance with the Finally, the sentences are ranked according to their scores, and the top N sentence is selected as the summary sentence.

The similarity of two sentences is calculated by comparing the number of identical words between the two sentences. Where Si and Sj denote two sentences respectively. The sentence similarity formula is shown below. Similarity(Si,Sj)=|WK|WKSiWKSjlog(|Si|)+log(|Sj|)$$Similarity\left( {{S_i},{S_j}} \right) = \frac{{\left| {{W_K}} \right|{W_K} \in {S_i} \cap {W_K} \in {S_j}}}{{\log \left( {\left| {{S_i}} \right|} \right) + \log \left( {\left| {{S_j}} \right|} \right)}}$$

The score of each sentence was calculated iteratively according to the following formula. The scores of the sentences were sorted in descending order, and the first N sentences with a high degree of importance were taken as the summary. WS(Vi)=(1d)+d*(ViIn(Vi)wjiVxout(vj)wjkWS(vj))$$W{S_{\left( {{V_i}} \right)}} = \left( {1 - d} \right) + d^*\left( {\sum\limits_{{V_i} \in In\left( {{V_i}} \right)} {\frac{{{w_{ji}}}}{{\sum\limits_{{V_x} \in out\left( {{v_j}} \right)} {{w_{jk}}} }}W{S_{\left( {vj} \right)}}} } \right)$$

By calling the TextRank algorithm to extract the text summary of each sub-topic, you can get the main content of the topic. For example, by calling TextRank algorithm on the text under the topic of “memory”, we can get the text summary as “So the second major component is the memory, which is also used to store programs and data in the computer. Then the memory also contains two major parts, one is the internal memory and the other is the external memory. U Disks are more compact and easy to carry around.”

Combining the TF-IDF algorithm and TextRank algorithm, we can simply take the overlapping part of the extracted summary content as the sub text summary. With the development of natural language understanding, text summary extraction algorithms are also constantly innovating and developing, and in the future about the text summary extraction part of the research task of generating mind maps, we can use more advanced algorithms to further improve the accuracy of the text summary extraction, and to provide a higher quality for the generation of mind maps.

Instructional design based on generative mind mapping

In order to implement the requirements of the “Information Technology” course standard for preschool education majors, improve students’ learning efficiency, and cultivate students’ active inquiry ability in the digital learning environment. To design preschool classroom teaching based on the mind map generated by teaching video proposed above. Following the principles of visualization, inspiration, induction, subjectivity and differentiation in the process of instructional design, using instructional video as a medium, generating mind maps as a scaffold, and mind maps as a tool, instructional design is conceived based on these three perspectives, and the roles of the instructional video, the generating mind maps, and the mind map tool are fully brought into play. In the instructional design process, the preschool education program is designed according to the three stages of pre-course preparation, in-class learning and post-course consolidation.

Knowledge Graph-based Personalized Recommendation Method for Teaching Resources
Contextualization methods based on resource maps

In order to effectively utilize the knowledge graph of teaching resources to accomplish the recommendation task, this paper utilizes the context in the knowledge graph of teaching resources from two aspects: local context information as well as non-local context information. In the pedagogical resource knowledge graph context processing method, the pedagogical resource knowledge graph is utilized to extract features and embed information for all learner history interaction items. Each history interaction item is regarded as a center node (head entity), and the global context embedding of each head node is obtained after processing its respective context information.

In the local context processing method, the first-order neighborhood entities of all historical interaction items are processed. The feature information of the first-order neighborhood entities of the head node is extracted first, and the relationship between the head node and the first-order neighborhood entities is then embedded. Then a join operation is performed on the two, and a learner-specific attention coefficient is computed by combining the head node information and the learner’s own characteristics. Next, using this attention coefficient, the embedding of each first-order neighborhood entity is linearly combined to obtain the local neighborhood embedding of the head node. Finally, the local context embeddings of all historical interaction items are obtained by aggregating the embeddings of the head node entities and the local neighborhood embeddings.

In the non-local context processing method, the non-local context of the head node is first extracted using a biased random wandering sampling strategy, then the gated loop unit is input in reverse order to output the non-local neighborhood embeddings, and finally the embeddings of the head node entities are aggregated with the non-local neighborhood embeddings to obtain the non-local context embeddings of all the historical interaction items.

The local and non-local context embeddings are fed into a gating mechanism that adaptively aggregates these two embeddings to finally obtain the global context embedding of the head node.

Description of local context processing methods

An entity in the knowledge graph is always linked to numerous other entities that can enhance its information in the knowledge graph. Graph neural networks can be used to gather feature information from neighbouring items in the knowledge graph, but aggregating neighboring information directly cannot capture learners’ specific preferences for entities. In order to identify learners’ personalized preferences for local contexts, this paper proposes a graph attention mechanism for each learner to aggregate the neighbor information of entities in the knowledge graph of teaching resources. For each learner, different attention coefficients are assigned to the same neighborhood entity of an item, and then these attention coefficients are used to aggregate neighborhood entities.

In this paper, we use Chl={t|(h,r,t)D}$$C_h^l = \{ t|(h,r,t) \in D\}$$ to denote the local neighborhood of entity h and define Chl$$C_h^l$$ as the local graph context of h in the knowledge graph of teaching resources. This study argues that inter-entity relationships in the knowledge graph of teaching resources play an important role in understanding semantic information, and if neighboring entities are connected through different relationships, they may have different impacts. In order to incorporate the relations into the attention mechanism, the embedding of the neighboring entities tChl$$t \in C_h^l$$ and the embedding of the corresponding relations r are first integrated by a linear transformation of Eq. (5): ert=(eret)W0$${e_{rt}} = \left( {{e_r}\left\| {{e_t}} \right.} \right){W_0}$$

where ∥ denotes the splicing operation and W02d×d$${W_0} \in {\mathbb{R}^{2d \times d}}$$ is the weight matrix. For the target learner u, the learner-specific attention coefficient describing the importance of entity t to entity h is calculated as shown in Equation (6): αu(h,r,t)=exp[πu(h,r,t)](h,r˜,x˜)=Dexp[πu(h,r˜,t˜)]$${\alpha _u}(h,r,t) = \frac{{\exp \left[ {{\pi _u}(h,r,t)} \right]}}{{\sum\limits_{(h,\tilde r,\tilde x) = \mathcal{D}} {\exp } \left[ {{\pi _u}(h,\tilde r,\tilde t)} \right]}}$$

Operation πu(h, r, t) is performed by a single-layer feed-forward neural network, which is defined as shown in Equation (7): πu(h,r,t)=tanh[(ehert)W1+b1]muT$${\pi _u}(h,r,t) = \tanh \left[ {\left( {{e_h}\left\| {{e_{rt}}} \right.} \right){W_1} + {b_1}} \right]m_u^T$$

where W12d×d$${W_1} \in {\mathbb{R}^{2d \times d}}$$ and b11×d$${b_1} \in {\mathbb{R}^{1 \times d}}$$ are the weight matrix and deviation vector, respectively. mu is a nonlinear transformation of eu which increases the flexibility of the model and it is defined as shown in Eq. (8): mu=ReLU(euW1˜+b˜1)$${m_u} = \operatorname{Re} LU\left( {{e_u}\widetilde {{W_1}} + {{\tilde b}_1}} \right)$$

where W1˜1×d$$\widetilde {{W_1}} \in {\mathbb{R}^{d \times d}}$$ and b˜11×d$${\tilde b_1} \in {\mathbb{R}^{1 \times d}}$$ are the weight matrix and bias vector, respectively. According to Eqs. (5) to (8), the attention mechanism becomes learner-specific due to the use of eu in the computation of the attention coefficients, which can reflect the learner’s personalized preference for neighbors. Given the normalization coefficients of each neighboring entity of h, linear aggregation of feature embeddings is performed on the entities in the local neighborhood to obtain the local neighborhood embedding of each head node as shown in Eq. (9): echl=tChlαu(h,r,t)et$${e_{c_h^l}} = \sum\limits_{t \in C_h^l} {{\alpha _u}} (h,r,t){e_t}$$

Then, the embedding of the aggregated entity h and its local neighborhood embedding echl$${e_{c_h^l}}$$ are aggregated to form the local context embedding cht$$c_h^t$$ of h, as shown in Equation (10): chl=tanh[(ehechl)W2+b2]$$c_h^l = \tanh \left[ {\left( {{e_h}\left\| {{e_{c_h^l}}} \right.} \right){W_2} + {b_2}} \right]$$

where W22d×d$${W_2} \in {\mathbb{R}^{2d \times d}}$$ and b21×d$${b_2} \in {\mathbb{R}^{1 \times d}}$$ are the weight matrix and deviation vector, respectively.

Personalized Recommendations by Fusing Behavioral Graphs

The recommendation task is to predict whether learner u is potentially interested in item i, which he has not interacted with before, based on the learner’s historical session information and the knowledge graph of instructional resources. Specifically, the ultimate goal is to learn the scoring function y^i$${\hat y_i}$$, which represents the probability that u may interact with item i, and based on the learned scoring function, the prediction scores of the candidate items are computed and ranked, and the top-ranked items are selected to be recommended to the learner. For the teaching resource recommendation task, this section proposes the DB-CGAT model, which combines the proposed teaching resource knowledge graph contextual processing method with the dual behavioral aggregation method, which is jointly used for multidimensional preference personalized recommendation of teaching resources.The structure of the DB-CGAT model is shown in Fig. 3.

Figure 3.

The DB-CGAT model

DB-CGAT consists of two main modules: a context processing module using resource graphs and a dual behavior aggregation module using behavior graphs. Based on the item representations and session representations learned from these two modules, the probability of a learner interacting with a target item is predicted for personalized multidimensional preference recommendations.

In the context processing module, the teaching resource knowledge graph is first processed with the teaching resource knowledge graph context processing method, where all the historical interaction items in the session are treated as the center nodes respectively, the feature information of their first-order neighborhoods is aggregated to the center nodes, and the embedding of the center nodes themselves is aggregated with the feature embeddings of the local neighborhoods to obtain the local context embeddings of the nodes. Next, the non-local context of the center node is extracted by biased random wandering sampling, and the non-local context embedding of the center node is obtained by using the gated loop unit. Finally, the local and non-local context embeddings are adaptively aggregated to obtain the global context embedding for each center node.

In the dual-behavior aggregation module, the node embedding representations containing structural information are involved in the operation of the dual-behavior aggregation method as input data, and all historical interaction nodes are added to the target behavior sequence and auxiliary behavior sequence, respectively. Through graph attention network and graph neural network learning, the structural global context embeddings of nodes are aggregated with the neighbor embeddings in the interaction sequences, and the embedding representation of each node is updated to obtain the true global representation of the node. The node global representations are substituted back into the behavioral sequences, and the target and auxiliary behavioral sequence representations are aggregated again using a gating mechanism to obtain a session-based representation of the learner’s preferences.

When performing multidimensional preference personalized recommendation, the initial embedding of the learner and the session-based learner representation are aggregated to obtain the learner’s contextual representation, which is again aggregated with the contextual representation of the target item to obtain the learner’s preference score of the target item, and based on the score the item to be recommended for the learner is determined in the end.

Effectiveness of AIGC technology integration applications
An empirical study of mind mapping-based teaching and learning

In this paper, the students of preschool education in a school were selected as the research object, and a questionnaire survey was conducted on two of the classes after the teaching of information technology based on mind mapping. A total of 183 students were surveyed, and the students in one of the classes at the performance point were set as survey group 1 and the students in the other class were set as survey group 2.

The questionnaire was designed to find out the students’ favorability and adaptability to the teacher’s change in teaching methods, and the statistics of the students’ responses in Survey Group 1 and Survey Group 2 are shown in Figure 4.

Figure 4.

Students’ fondness for the new teaching method

The statistics show that more than half of the students like the new way of teaching, especially in survey group 1 up to nine out of ten students like it, more than one-fifth of the students like the new way of teaching very much, and about one-tenth of the students have an indifferent attitude. More than one-half of the students in Survey Group 2 “liked” the new method of instruction very much, nearly nine-tenths liked it, and less than one-tenth of the students were “indifferent” to the new method of instruction. While zero students in Survey Group 1 chose “dislike” and “strongly dislike”, about one in twenty students in Survey Group 2 chose “strongly dislike”. Overall, most of the students in this program “like” the new teaching method and they have a positive attitude towards IT teaching based on mind mapping.

The students’ statistics on the appropriateness of the teaching style for their learning are shown in Figure 5.

Figure 5.

Students’ adaptability to the new teaching methods

As can be seen from the figure, most of the students think that the new teaching method is suitable for them, especially in survey group 1, 90 percent of the students think that the teacher’s teaching method is suitable for them, and in survey group 2, 70 percent of the students think that the new teaching method is suitable for them. About one-fifth of the students in Survey Group 2 felt that they “couldn’t tell” whether the teacher’s teaching style was suitable for them, and one-tenth felt that the new teaching style was not suitable for them. Less than one-tenth of the students in Survey Group 1 thought that they could not tell whether the new teaching style was suitable for them, and there were also some students who thought that the new teaching style was not suitable for them.

As can be judged from Figures 4 and 5, more than four-fifths of the students like the new teaching method and think that the new teaching method suits them, which verifies the effectiveness of the teaching method based on mind mapping. At the same time, a very small number of students chose dislike and unsuitability, indicating that students can have more subjective judgments when dealing with the new teaching, showing that the results of this survey have authenticity and accuracy.

Effectiveness of personalized recommendation methods for teaching resources
Data sets

To validate the effectiveness of the DB-CGAT model for recommendation tasks, the experiments in this chapter are conducted using the publicly available Yelp2018, Amazon-Book dataset and the self-built CoLR dataset. These datasets differ in terms of domain, data size, and sparsity. The following is a brief description of the datasets:

Yelp2018 dataset for business location recommendations.Yelp2018 is a dataset provided by Yelp, the largest social consumer review site within the U.S. Yelp2018 features addresses, ratings, review content, and user information for registered businesses across the U.S. in 2018, and contains data from more than 6,000,000 user reviews covering all 50 U.S. states.

The Amazon-Book dataset is used for book product recommendations. AmazonBook collects metadata on over 96,000 books on the Amazon shopping site, which includes information such as title, ISBN code, author, publisher, publication date, user reviews, and category. The data size is about 92,000 rows with 13 fields per row, which includes information such as book title, ISBN code, author, publisher, user reviews and category.

CoLR dataset is used for recommendation of teaching resources.CoLR is the teaching resources dataset constructed in this paper, which collects information about course content, instructor information and lab support from several higher education teaching websites with a data size of about 31,000 items, mainly for the fields of computer and electronic information.

Each dataset consists of two parts: user-item interaction graph and knowledge graph, in addition to user-item interactions, a corresponding knowledge graph needs to be constructed for each dataset. Referring to a similar setup in KGAT, the knowledge graphs for the three datasets are constructed by mapping items to Freebase entities through title matching, extracting the triples that are directly related to the item-aligned entities and considering nodes in their two-hop range. Various types of information are sampled in the datasets as entities of the knowledge graph, such as category and location of business premises, author and publisher of books, author and type of instructional resources, etc., and connectivity relationships between entities are generated.

Experimental results and analysis

Model recommendation performance experiment

The results of the tests on the accuracy of the recommendation tasks are shown in Table 1. It can be found that DB-CGAT outperforms other baselines and brings significant improvement in recommendation results on all three datasets, which validates the effectiveness of the knowledge graph-based DB-CGAT model for recommendation tasks. The datasets used for the evaluation produce different result preferences due to the variability of factors such as data sparsity, knowledge graph features, and different recommendation scenarios. The results in the table show that DB-CGAT can achieve better performance in most cases in the comparison of the different datasets listed and several mainstream baseline methods, which proves the versatility and flexibility of the proposed DB-CGAT model. Benefiting from the combination of the contextual processing method and the dual behavioral aggregation method, DB-CGAT can denoise entity-related relationships and capture more accurate semantic features.

Experiments on the effect of mitigating data noise

To verify the usefulness of DB-CGAT for recommender systems in knowledge graphs with noisy data, tests are conducted on three datasets, Yelp2018, Amazon Book, and CoLR. Table 2 shows the experimental results of processing noisy data by inserting noisy triples and random connections into the knowledge graph to simulate knowledge graph noise caused by irrelevant entities.

Test results of baseline model

Model Yelp2018 Amazon-Book CoLR
Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20
BPR 0.0368 0.0353 0.1253 0.0636 0.0682 0.0331
LightGCN 0.0562 0.0252 0.1264 0.0639 0.0726 0.0389
CKE 0.0532 0.0352 0.1276 0.0621 0.0692 0.0428
RippleNet 0.0672 0.0423 0.1368 0.0683 0.0621 0.0320
KGAT 0.0662 0.0438 0.1085 0.0714 0.0767 0.0311
CKAN 0.0529 0.0373 0.1321 0.0572 0.0813 0.0479
DB-CGAT 0.0732 0.0472 0.1396 0.0709 0.0897 0.0518

Experimental results of processing noise data

Model Yelp2018 Amazon-Book CoLR
Recall@20 NDCG@20 Recall@20 NDCG@20 Recall@20 NDCG@20
KGAT 0.0664 0.0453 0.1032 0.0696 0.0726 0.0347
CKAN 0.0563 0.0426 0.1226 0.0517 0.0742 0.0393
DB-CGAT 0.0791 0.0584 0.1335 0.0783 0.0847 0.0501

From the results presented in Table 2, it is clear that the proposed DB-CGAT model still has better performance when competing with other knowledge-aware recommendation models. Specifically, DB-CGAT works best in mitigating knowledge graph noise, while still generating the best evaluation results on items with sparse knowledge entities.

Experiments on the effect of the temperature parameter τ.

In the DB-CGAT model for contextual processing, due to the existence of certain negative samples that are very similar to the feature representation of the positive samples leading to the model is difficult to separate these difficult negative samples from the positive samples, so the identification of the difficult negative samples contributes a lot to the gradient descent of the loss function, and it can effectively speed up the convergence of the model results and generate a more robust parameter of the node feature representation τ Mining of difficult negative samples has an important significance for mining difficult negative samples. The performance of different τ-value models under different datasets is shown in Fig. 6.

Figure 6.

Performance of models with different tau values under different data sets

In both CoLR and Yelp2018 datasets, the DB-CGAT model has the best Recall@20 recommendation performance when τ = 0.3. And when the value of τ is too large, it will lead to poor performance and require more rounds of epoch training to converge, while too small a value of τ will harm the model performance due to convergence and other issues. Therefore, the choice of τ should not be too large and too small, and the τ value can be fine-tuned in the range of [0.1,1.0] to determine the best choice of τ value.

Conclusion

This paper takes the teaching of information technology in preschool education as an example to study the integration application of AIGC technology-enabled education automation.

According to the data collected by the questionnaire method, it is obtained that more than 80% of the students like the teaching method based on the mind map and think that the new teaching method is suitable for them, which verifies the effectiveness of the teaching method based on the mind map.

The recommendation results of DB-CGAT model and six mainstream baseline methods on three datasets show that DB-CGAT can achieve better performance in most cases. On the CoLR dataset, the Recall@20 accuracy of the DB-CGAT model was 0.0121 and 0.0105 higher than that of the KGAT and CKAN models, respectively, and the NDCG@20 accuracy was 0.0154 and 0.0108 higher than that of the KGAT and CKAN models, respectively. In the Yelp2018, Amazon Book datasets, the DB-CGAT model also works best in mitigating knowledge graph noise. In both CoLR and Yelp2018 datasets, DB-CGAT model has the best Recall@20 recommendation performance when τ = 0.3.

Sprache:
Englisch
Zeitrahmen der Veröffentlichung:
1 Hefte pro Jahr
Fachgebiete der Zeitschrift:
Biologie, Biologie, andere, Mathematik, Angewandte Mathematik, Mathematik, Allgemeines, Physik, Physik, andere