Open Access

Improving Mandarin Emotional Vocabulary in Application Writing Using Deep Learning Models

  
Sep 26, 2025

Cite
Download Cover

Introduction

Practical literature, as the name suggests, is a “text” chapter that “should” be used in life and practice, which has the characteristics of authenticity and practicality, and is a kind of literary style formed by human beings in long-term social practice activities, and is a tool for people to transmit information, handle affairs, and exchange feelings, and some applied texts are also used as evidence and basis [1-2]. With the development of society, people’s interactions in work and life are more and more frequent and things are more and more complicated, so the function of application essays is more and more. Discursive, narrative and epistolary application essays usually contain part of the emotional vocabulary, which can effectively stimulate the emotional resonance of readers and attract their attention, and different emotional vocabulary can produce different effects. With the popularization of the Internet and social media, people use more and more emotional words in daily communication. These words have enriched people’s understanding and ability to express emotion. However, in practical application, people usually use emotional vocabulary with various problems, such as ambiguity of meaning between different words and confusion of emotional polarity [3-4], which causes inaccuracy in the application of emotional vocabulary in application essay writing.

In the past few years, deep learning techniques have been widely used in sentiment analysis. Deep learning is a machine learning technique that uses multi-layer neural networks to learn features and perform classification [5]. Text categorization is the core task of sentiment analysis, which classifies text as positive, negative or neutral [6]. Deep learning can perform text classification by learning the features of text. Specifically, deep learning can use convolutional neural network or recurrent neural network to learn the features of text and use softmax classifier to categorize text into different classes [7]. Sentiment lexicon is an important part of sentiment analysis. Traditional lexicon-based sentiment analysis methods use sentiment vocabularies to identify the sentiment of a text. However, there are some problems with this approach, such as the limited coverage of sentiment vocabularies, which cannot satisfy all the needs of sentiment analysis [8-9]. Deep learning can extend sentiment vocabularies by using a large amount of text data. Specifically, deep learning can use word vectors to represent each word and use text categorization methods to learn new sentiment words. In this way, deep learning can extend the existing sentiment vocabularies and improve the accuracy of sentiment analysis [10-12]. Traditional sentiment analysis methods usually can only identify the positive, negative or neutral sentiment of a text, and cannot analyze the specific details of the sentiment. Deep learning can use recurrent neural networks to identify the diversity of the sentiment, learn the hierarchical structure in the text, and take into account its contextual information when identifying the sentiment [13]. Such an approach to sentiment analysis undoubtedly brings new opportunities for the accuracy of vocabulary usage in applied texts.

In this study, we first propose a method for constructing an emotion dictionary based on the text of application essays, using certain rules to clean and filter the data, and propose an average scoring mechanism using user ratings to determine the emotion polarity of the target words, and at the same time, calculate the emotion intensity of the target words, and finally construct an emotion dictionary based on the writing of application essays. Then the method of constructing a bi-directional LSTM-based sentiment dictionary for Mandarin words is introduced, which improves the CBOW model into an ECBOW model and adds part of the network structure to get the sentiment features, and then describes the semantic dependencies according to a Huffman semantic binary tree, and uses a bi-directional LSTM neural network to learn the features for the bifurcation of each word’s bifurcation of the semantic dependency path information, and obtains the bifurcation of semantic paths features, and finally, the distance feature from the word to the center word and the center word information feature are added as the overall features of the word and input into the neural network, and supervised neural network training is carried out to obtain the dynamic sentiment lexicon. Subsequently, lexicon evaluation experiments are conducted using precision, recall, and F1 metrics with the general generic sentiment lexicon. Finally, the proposed bi-directional LSTM-based dynamic sentiment lexicon was applied to sentiment mine application writing texts from learners in five types of institutions to verify the feasibility and effectiveness of the proposed method.

Methodological design for the construction of an emotional lexicon

This emotion dictionary construction method can be divided into the following 3 steps: text pre-processing, i.e., the movie review data should be pre-processed through a series of text processing; judgment of word emotion polarity based on the average rating mechanism, i.e., the polarity of the word is judged through a certain method by fusing the user ratings and the number of likes; computation of word emotion intensity based on the average rating mechanism, i.e., the intensity of the emotion is computed through the word emotion polarity in the previous step.

Text pre-processing

After text data collection, preliminary text processing is required before subsequent affective tendency calculations can be performed.

Mandarin affective lexical disambiguation

It is generally believed that in ensuring efficiency at the same time, jieba participle is a better participle effect of the participle tool, so the choice of jieba participle tool for participle. At present, jieba has a total of four kinds of word separation, respectively, the precise mode, full mode, search engine mode and paddle mode. Precise mode refers to the sentence by default method in order to seek the most accurate cut, for the movie review data, more suitable for the use of this mode.

Removal of deactivated words, low-frequency words

Deactivated words often refer to a type of word that occurs more frequently in a text, but does not have an impact on the information contained in the text. For sentiment analysis, these words are also usually neutral, so they can be removed first. First, the Baidu deactivation word list was chosen for the text of the application because it works best for most corpora on average. Second, a small proportion of the deactivated words in the word list were found to overlap with words in the Mandarin Emotion Vocabulary Ontology. Finally, these overlapping words were excluded from the deactivation word list to avoid affecting the subsequent calculations.

Removing symbols and expressions

Symbols or emoticons in text carry some sentiment information though, and they are also used in sentiment analysis studies. However, since the existing manually labeled sentiment lexicons generally do not have entries for symbols or emoticons, this information is not considered in this study for the time being and needs to be removed all together.

Judgment of word sentiment polarity based on average rating mechanism

The PMI algorithm is often used to determine the degree of association between words. As mentioned before, early scholars would manually select some very obvious positive and negative emotion words as standard emotion words, and use the PMI algorithm to find out the words with higher degree of correlation with these standard emotion words in the corpus to expand the emotion lexicon. After that, many scholars tried to utilize the improved PMI algorithm to expand the sentiment lexicon, and their approaches were largely the same.

A large corpus of application writing was first built and gradually narrowed down the set of candidate sentiment words to be constructed. First, the user’s ratings of a minimum of 1 and a maximum of 5 are utilized to classify positive and negative words, setting 4 and 5 as positive comments, 1 and 2 as negative comments, and discarding comments with 3 points. Subsequently, the number of times the comment where the candidate vocabulary is located is set as positive or negative is counted, and the PMI algorithm is modified to compute the mutual information between the candidate vocabulary and the positive or negative comment.

First, generalizing the scores of 4 and 5 as positive or 1 and 2 as negative loses information about the degree of user evaluation, which may have an impact on the judgment of the word’s sentiment polarity. Second, discarding the 3-point reviews, which represent neutrality, wastes the relevant dataset to some extent. Therefore, this study proposes an average rating mechanism that, first, considers user ratings. Secondly, the number of likes on user comments by other users is considered. While the level of user ratings can certainly determine the emotional tendency of a candidate, the number of likes on it by other users can also reflect whether the comment is helpful to other users, reflecting the credibility of the content of the comment. The scoring mechanism is shown in equations (1) to (4). Sum(t)=i=DSi

where is used to calculate the total rating of candidate word t (D refers to the corpus) and only user ratings are taken into account without considering the number of user likes. St refers to the user rating of the comment where candidate word t is located, and its value range is [1, 5]. Sum(t)=tDSt+{ logαLt(St>3) 0(St=3) logαLt(St<3)

where the total rating of candidate t is calculated, taking into account the user ratings as well as the number of user likes. Lt refers to the value of the likes of the comment where candidate word t is located, and the base of the logarithm a is generally taken as [10, 100], which can be subsequently adjusted depending on the accuracy of the experiment. Avg(t)=Sum(t)Count(t)

where the average rating of candidate t was computed for a range of [1, 5]. Scr(t)=Avg(t)t=DAvg(t)t=DCount(t)

where the final score of candidate word t can be calculated. If the final score is positive, candidate word t is categorized as a positive sentiment word. If the final score is negative, it is categorized as a negative sentiment word.

Calculation of word sentiment intensity based on average rating mechanism

Normalization is performed at this point so that the final score falls within the interval [-1, 1] for practical applications. Normalization method calculation: NScr(t)={ Scr(t)minPmaxPminP(Scr(t)>0) Scr(t)minNmaxNminN1(Scr(t)<0)

Where, maxp and minp denote the maximum and minimum values when the candidate word sentiment score is positive, and similarly, maxx and minx denote the maximum and minimum values when the candidate word sentiment score is negative.

Bidirectional LSTM-based dynamic sentiment dictionary construction method

In the text sentiment classification based on the sentiment dictionary, the construction of the sentiment dictionary is a key link in the sentiment classification, and the quality of the sentiment dictionary often affects the accuracy of the sentiment classification, and a high-quality sentiment dictionary can provide more comprehensive information about the sentiment, thus improving the quality of the text sentiment analysis [14-15]. The process of constructing the sentiment dictionary is shown in Figure 1.

Figure 1.

Emotion dictionary construction process diagram

Data and basic processing

Emotion words refer to words with emotional color in the text, this paper stipulates that emotion words are composed of emotion word body and sub-body, and its sub-body can be empty, and stipulates that emotion words are classified into positive emotion words, negative emotion words, and neutral emotion words. The classification of emotion words is shown in Figure 2.

Figure 2.

Classification of emotion words

Emotional feature extraction of vocabulary
Word Vector Learning Using the CBOW Model

CBOW model is a model that predicts the occurrence probability of the current word based on the words in the context. The ECBOW model diagram is shown in Fig. 3. The left half represents the CBOW model, whose neural network structure is divided into input layer, projection layer and softmax layer, and the vocabulary w represents the one-hot vector, and V represents the m×|V| word vector table [16]. After the conversion of the word vector table into the projection layer, the vocabulary vectors vi are superimposed as shown in Equation (6). vt=ivi,i{tm,....,t+m}

Figure 3.

ECBOW model diagram

In this paper, we set m = 2. The maximum probability of w is output at the output layer, and then the maximum likelihood is normalized to the function, as shown in Equation (7). fw=1|c|i|c|logp(w|wim,wi(m1),...,wi+m)

Extracting Sentiment Features Using ECBOW Modeling

The ECBOW model proposed in this paper is based on the principle that based on the original CBOW model that constrains the syntactic structure of the context, the added emotion constraint network structure constrains the emotion information of the text with positive and negative emotions [17]. In the added network structure, the input layer is the word w in the text with positive and negative polarity, and the projection layer sj represents the one-hot vectors of all the words in the text with positive and negative meanings obtained by transforming them into word vectors and summing them through the word vector table v, as shown in Equation (8). sj=invi,i{ 1,2,....,n}

The text is categorized into tweets with positive and negative sentiments, so a logical neuron is utilized to output the probability of being positive and negative as shown in Equation (9). P(hj|sj)={ σ(sjH)p(sj)=[1,0] 1σ(sjH)p(sj)=[0,1]

Where H is the vector parameter and notation p(sj) is the text praxis, if the projection layer sj is projected by praxis text then p(sj) = [1, 0]. If the projection layer sj is projected by praxis text then p(sj) = [0, 1]. Thus maximizing the objective function as shown in equation (10). ζ=α1|s|j=1|s|logP(hj|sj)+(1α)fw

Where S denotes the set of sentences in the dataset, θ parameters include U, V, α, η, U for the left half of the projection layer to softmax layer vector parameters. α denotes the weight parameter. η denotes the learning rate. e denotes the magnitude of vector change. Where: V(w) denotes the vector of vocabulary w. xs=wesjV(w) hj={ 1 p(sj)=[1,0] 0 p(sj)=[0,1]

Bifurcated Semantic Dependency Path Feature Extraction
Huffman Semantic Dependent Binary Trees

Huffman semantic dependency binary tree is described by lexical nodes and lexical dependencies and Huffman encoding, Huffman semantic dependency structure binary tree construction is generated based on the semantic dependency tree structure, the semantic dependency structure binary tree diagram is shown in Figure 4.

Figure 4.

Binary tree of semantic dependency structure

Cross-Semantic Dependency Path Representation

Since the center word can usually represent the main syntactic and semantic features of the phrase, it is considered to have strong predictive ability. In this paper, the root node of Huffman bifurcation semantic dependency structure is specified as the center word, and the path from words to the center word in the text is called bifurcation semantic dependency path in this paper.

Extracting Bifurcated Semantic Dependency Path Features

In this paper, with the help of language modeling, the bifurcated semantic dependency path of each word in the text is taken as a text sequence, and the product of the probability of occurrence of the bifurcated semantic dependency path of all the words in the whole sentence, i.e., PL(s), is computed, as shown in Eq. (13). PL(s)=i=1sPL(wi)

where wi is a leaf node in a semantic dependency binary tree. L denotes the sequence of paths of wi. PL(wi) is the probability of occurrence of the binary semantic path of this node wi. The bifurcation semantic dependency path feature extraction using bidirectional LSTM neural network is shown in Fig. 5. The input layer is the node in the bifurcation semantic dependency path of each word. The bifurcation semantic path feature acquisition is carried out through the bidirectional LSTM neural network, whose feature vector is fp, and whose dimensionality is np. The neural network is finally made to converge by maximizing the objective function as shown in Eqn. (14). The feature acquisition is carried out through the bidirectional LSTM neural network. The vector representation of the feature layer i.e. the bifurcated semantic dependency path feature for each word in the text. f=1Ni=1Nlog(PL(wi))

Figure 5.

Extraction of binary semantic dependency path feature map

Relative position and sentence center word information

In the utterance, the center word has a strong predictive role, it can skip some auxiliary or adverbial thus achieving long-distance structural constraints, usually the center word can represent the main syntactic and semantic features of the phrase. Relative position indicates the position of words in the utterance relative to the center word. In this paper, the relative position feature is mapped into a nrp-dimensional vector frp.

Sentiment Dictionary Classifier Construction

In this paper, we view the sentiment dictionary classifier construction problem as a sequence labeling problem, for each word in the text contains four features respectively: sentiment word feature fe, bifurcated semantic dependency path feature fp, center word information fc and relative position feature fp. The final synthesized feature f0 is denoted as: f0=[fer,fpr,fcr,frpr]r

Remember that the vector dimension of emotion word feature fϵ is de, the vector dimension of bifurcated semantic dependency path feature fp is dp, the vector dimension of center word information fc is dc, and the vector dimension of relative position feature frp is drp. Therefore, the dimension of the final combined vector fθ is dim(f0) = dc + dp + dc + drp, and in this paper, we use a bidirectional LSTM network to process the sentence, so as to obtain the forward representation of the sentence h , and the inverse representation of the sentence h , and choose tanh as the activation function to produce the final representation of the sentence as shown in Equation (16). The construction of emotion word classifier graph using bidirectional LSTM neural network is shown in Fig. 6. h=tanh(h+h)

Figure 6.

Constructing emotion word classifier using bidirectional LSTM neural network

Text labeling framework

In this paper, the emotional lexical features are extracted through the text labeling framework, and this paper defines the PNO framework as the text labeling framework, noting the frame label T ∈ {P, N, O}, where the frame labels are denoted as follows: if there is a word with positive emotional color in the text, this paper labels the word as P: if there is a word with pejorative emotional color in the text, this paper labels the word as N: if there are words with positive emotional color and pejorative emotional color neither reflected in the text, this paper labels the word as O. color and pejorative emotional color are not reflected, this paper marks the word as O.

Output layer representation

The output layer is calculated as shown in Equation (17). O=softmax(Wsh+bs)

The dimension of the output node in the output layer is 3, with [1, 0, 0], [0, 0, 1], [0, 1, 0] to indicate that the output result of the term is positive (P), neutral (O), and negative (N), respectively. Then it is processed by softmax layer to find out the probability of belonging to a certain category, as shown in equation (18). p(xi)=exp(xi)jexp(xj)

In this paper, the loss function cross entropy is used as the objective function, as shown in equation (19). L=iiyiilog(preii)

Where, yn refers to the ird value in the actual label at the t moment and preε refers to the ith value in the predicted label at the i moment. Where: e(w) ∈ D, the bidirectional LSTM sentiment classifier is first trained to complete the construction of the dynamic sentiment dictionary, and then the test text set is predicted by this sentiment classifier, if the output is P-labeled, then the current vocabulary is a positive-sense sentiment word and the positive-sense sentiment word is deposited into the positive-sense lexicon Dpos. If the output is N-labeled, then the current vocabulary is a negative-sense sentiment word and the negative-sense sentiment word is deposited into the negative-sense lexicon Dneg. If the output is O-labeled, then the current vocabulary is neutral and finally the primary extended static sentiment lexicon is obtained.

Experimental results and analysis
Emotional vocabulary recognition experiment

The experimental analysis of this paper includes two parts, data acquisition and description, and evaluation of experimental effect. In the data acquisition part, the data of short user reviews from the largest application essay writing website is used as the sample of the sentiment dictionary construction experiment. The experimental effect evaluation is measured using precision rate, recall rate, and F1 value.

Data acquisition and description

Firstly, the most popular social group in the website was selected, then the IDs of the users in the community were collected, and finally 396,1207 users’ IDs were collected, for which, after de-emphasizing the object of the comment by using the ISBN number, a total of 12,118,836 applications were collected, and an average of 3.3 applications were commented by each user. For each comment, fields are captured including “user ID”, “rating”, “comment content”, “web link”. In this paper, we mainly consider the construction of Mandarin sentiment lexicon, so we exclude comments that do not contain Chinese characters, comments that are a mix of Chinese and English, and comments that contain only emoticons as well as gibberish, leaving a total of 2313436 comments. After removing the three-star ratings representing neutrality, the statistical information table of short comments in the word frequency statistics stage is shown in Table 1.

The statistical phase of the statistical phase of the word frequency

Typo 1 star 2 star 4 star 5 star Forward Negative direction Total quantity
Quantity/piece 61986 121715 973369 1226827 2285165 181825 4850887

Due to the experimental needs, the relevant fields of each comment data include “User ID”, “Rating”, “Comment”, “PageUrl” If the content of a field is empty, the comment is excluded. The fields of the final experimental corpus mainly include vocabulary and the distribution of vocabulary on ratings. Sample vocabulary rating statistics are shown in Table 2.

Vocabulary score statistics

Vocabulary Book number Number of users 1 star 2 star 3 star 4 star 5 star
Introduction 1207 1234 4 29 363 598 389
Good 1984 2217 21 57 370 1150 841
Unintelligible 1119 1315 250 349 359 281 165
…… 637 577 25 53 349 155 86
Look at it 814 1039 24 115 351 354 328
Incomprehension 1207 1234 -1 29 363 598 389
Experimental procedures

The experiment as a whole consists of two interrelated segments, lexicon filtering and sentiment score calculation. According to the rules and process of generating the candidate word set of the sentiment lexicon, the short comments in the database are extracted to generate a collection of ultrashort comments. First, among all 197,078,085 entries, preliminary filtering is carried out using the length of the comments, retaining the short comments with a length of 10 characters or less, and if the short comments appear repeatedly, they are counted as appearing only once, and the score is the score with the highest frequency, which yields 929,250 candidate emotion words. According to the calculation and filtering ideas in the previous section, the candidate emotion words are processed and the non-emotion words are eliminated. All the rules are formulated with a total of 9 rules. The rule set formulation and specific application in the model are as follows.

Candidate words with a total word frequency of less than 3 were eliminated. According to the observation of the data, the illustrative users of the words with too low frequency are fewer and not representative. After this cleaning work, the score of each candidate emotion word is calculated according to SO-PMI, if the subsequent culling work is carried out first, it will result in the loss of emotional information, according to the characteristics of the data, the threshold H is set to 2. Sample results of the emotion word calculation are shown in Table 3. The positive and negative frequency comparisons of candidate words show that users are very inclusive of resources, and when the threshold H of SO-PMI is set to 1, words such as “embarrassment” and “feel general” are still positive. However, when the threshold H is set to 3, it is too harsh, and the intensity of common positive words such as “healing” and “not bad” does not even reach 1. Therefore, taking the above considerations into account, the threshold is set to 2.

Eliminate candidates with an emotion score less than 0 and a negative frequency freq_n equal to 1.

Eliminate candidates for non-common nouns.

Eliminate candidates that only contain parts of speech “c”, “d”, “g”, “h”, “k”, “p”, “q”, “r”, “u”, and “x”.

Eliminate candidate words that do not contain parts of speech “a”, “b”, “d”, “n”, and “v”.

Eliminate candidates that only contain a single part of speech “n” and the absolute value of sentiment score |score|is less than 0.35.

Eliminate candidates with modal words such as “ah”, “ah”, and “bar”.

Candidate words containing one of the following characteristic words: “world”, “love”, “female”, “male”, “sir”, “high school”, “childhood”, “human nature”, “youth”, “electronic version”, “elementary school”, “university”, “life”, “life”, “textbook”, “high school one”, “person”. According to observations, these common nouns are more frequent in the field of practical writing, but they do not reflect emotional content and are deleted. After cleaning according to all the rules. Table 4 shows the record of the rule-based filter process of candidate sentiment words.

The emotional word calculates the sample

Candidate word Forward frequency Negative frequency Threshold value=1 Threshold value=2 Threshold value=3 Emotional intensity
Very General 34 43 -1.2438 -2.1507 -3.0974 -0.6727
Cure 846 18 2.8314 2.0781 1.1103 0.9815
Good 693 14 2.9508 1.5632 0.6642 0.9286
Generally 34 12 0.4187 -0.7638 -1.5323 -0.1298
Chicken Soup 30 61 -1.6404 -2.466 -3.535 -0.6016
Embarrassing 213 46 0.3502 -0.5849 -1.4594 0.6812
Dry Goods 33 27 -0.8731 -2.0351 -3.0566 -0.4729
Brain Hole 688 9 2.9753 2.2108 1.2001 1.1169

A conditional filtering process recording table based on rules

Elimination rule Remaining candidate affective words Positive word Negative term
Initial 929229 -- --
(1) 90307 69943 20330
(2) 77557 66979 10585
(3) 65684 61547 8153
(4)~(6) 62930 55498 7405
(7)~(9) 59383 52536 6804

After some processing operations, a sentiment lexicon based on large-scale ultra-short comments oriented to the field of application writing is formed, and the Top30 high-frequency sentiment words are shown in Table 5.

High frequency emotional term top30

Positive emotion Good, good, good, great, great, great love, great love, great love, good, still, move, a little meaning, a little meaning, the recommendation, great, the shock, is worth reading, is worth reading, touching, is worth seeing, good, warm, very interesting
Negative emotion Generally, I don’t understand, garbage, don’t like, don’t look, generally, don’t understand, see not to go, copy, still, I don’t know, I don’t read it, I don’t read it, I don’t know, it’s very general, no sense, bad, general, melodramatic, I don’t understand, I can’t read it. No meaning, tail, disappointment, Have seen, What the devil, Forget, forget, read and don’t understand, speechless
Evaluation of experimental effects

In order to evaluate the sentiment dictionary constructed in this paper, two indicators, recall rate and accuracy rate, are introduced and compared with other general dictionaries. Recall rate and accuracy rate are important evaluation indexes in the field of information retrieval and information recommendation, and they are also widely used for the quality evaluation of emotion dictionary. Define the confusion matrix for evaluating the experimental effect, and the confusion matrix for evaluating the experimental effect is shown in Table 6.

Experimental effect evaluation confusion matrix

Forecast
Correctness Errors
Actual Correctness TP FN
Errors FP TN

Table 7 shows the comparison results of sentiment dictionary pairs. By analyzing the words included in the dictionary constructed in this paper but not included in the other two general dictionaries in the recall sample set, it is found that the dictionary better contains the characteristic words in the field of applied writing, such as “illustrated and illustrated”, “typo”, etc. Dictionaries are better at discovering new words, such as “unclear”, “powerful”, etc. Dictionaries do a good job of including colloquial words, such as “read in one sitting”, “worth seeing”, etc. A dictionary can contain word ontologies and collocations, such as “pretty”, “too long”, etc.

The emotional dictionary is the same as the result

Categories Dictionary size Recall rate Accuracy rate F1
English language dictionary 59115 91.34% 78.48% 83.73%
Dalian polytechnic emotional dictionary 27463 37.54% 70.38% 48.09%
The hownet emotional dictionary 8733 35.34% 75.76% 46.92%
Emotion mining in application essay writing based on emotion lexicon

The textual material for this study was taken from the corpus of “Writing for the Millionth Applied Writing Competition”, which consists of a total of 734,529 writing texts. The writers of the texts in the corpus are from five different types of colleges and universities. Since the texts are written on the same topic, the possibility of differences in linguistic features and emotions of the texts due to factors other than the individual writers is low. The average length of texts from writers from the five types of institutions ranged from 130.33 to 220.08 words, and the average number of sentences per text ranged from 21.14 to 26.43 words. The statistics of writing text features are shown in Table 8.

Writing text feature statistics

Institutional type Text number Basic statistics Grammatical index
Total word The words are all words It’s greater than the ratio of 6 letters Proportional pronoun (%) Proportional pronoun (%) Proportional pronoun (%) Proportional pronoun (%)
Key university 86574 220.08 21.44 27.57 10.94 7.53 14.16 5.83
School of our school 602045 193.14 25.61 25.41 12.38 9.63 13.95 7.19
Vocational school 40405 188.85 26.43 30.35 7.41 7.25 13.74 9.16
High school 4042 140.66 23.66 22.58 9.67 6.14 14.62 3.99
Junior high school 1463 130.33 21.14 14.54 11.76 9.38 12.04 6.38
Emotional tone

The differences in the characteristics of writing texts from the five types of institutions for the four new composite statistical indicators (analytical thinking, impact, authenticity, and emotional tone) are shown in Figure 7. It can be found that the differences between the writing texts from the five types of institutions in these four comprehensive indicators are more obvious, especially in the two of authenticity and emotional tone, the writing texts from high schools are significantly higher than the other four types of institutions. In terms of the emotional tone of the text, the mean value of writing texts from middle schools was much lower than that of the other four types of institutions. The writing texts from key universities, secondary institutions, vocational institutions, and high schools all scored well above 75, showing a significant positive affect. In contrast, the overall emotional tone of middle school students’ writing texts had a mean of nearly 50, indicating that middle school students’ writing did not show a clear positive or negative emotional tendency. The standard deviations of each indicator for writing texts from all five types of institutions were large, suggesting that even individual writers from the same type of institution can vary widely in their performance on the same indicators.

Figure 7.

The characteristics of writing in five schools are different

Positive and negative emotions and their decomposition

The five types of institutional applications as positive and negative emotions in writing texts and the decomposition of negative emotions are shown in Figure 8 (Figure a represents emotional processes and Figure b represents negative emotions). By examining the two secondary categorical indicators of positive and negative emotions in the emotion lexicon, this study found that the frequency of words expressing positive emotions was much higher than the frequency of words expressing negative emotions in writing texts other than middle school. Taking positive emotion words as the statistical target, the frequency of positive emotion words appeared in the writing texts of high school students was the highest, while middle school was the lowest. The frequency of positive emotion words in the composition texts of key universities, second colleges and vocational colleges resided between junior and senior high school students, and there was no significant difference within the three. Among the negative emotions, the frequency of words in the texts of middle school students was the highest among the five types of schools and much higher than the other four types of schools.

Figure 8.

Positive, negative and negative to emotional decomposition

The top 20 positive affective lexical items that appeared most frequently in the writing texts of the five types of institutions are shown in Figure 9. In the high-frequency word use of positive emotion-related words, the various types of schools reflected certain common emotion lexical items, but each still had some unique high-frequency emotion lexical items. Taken together, the general positive emotion words such as good, well, better, great, best, and love were widely used in their own right and ranked high in frequency in students’ writing texts at all levels. Words commonly used to modify specific people or objects appeared more frequently in the texts of junior high and high school students’ compositions. For example, the word frequency of beautiful ranked 5th among the positive affective words used by junior high school students and 9th among the positive affective words used by high school students, both of which were relatively high, and the word did not appear among the top 20 high-frequency positive affective words in the texts of the students from key universities, second-base colleges and vocational colleges.

Figure 9.

The top 20 positive emotional entries in the writing text of five schools

The top 20 negative affective lexemes that appeared most frequently in the writing texts of the five types of institutions are shown in Figure 10. The top 20 high-frequency words used by middle school students were more clustered in the use of negative affective words in the essays of the five types of institutions. For example, the negative affective word broke, which ranked 20th in frequency in the middle school text, appeared much more frequently than words that ranked much higher in frequency in the essays of students from key universities, second colleges, and vocational schools, such as bad. In contrast, high school students’ use of negative affective vocabulary was relatively dispersed, with even the word DIFFICULT having a relative frequency of only slightly more than 0.022%.

Figure 10.

The top 20 negative emotional entries in the writing text of five schools

In order to further explore the affective differences and their causes among learners writing texts in different types of institutions, the categorical indicator of drive was selected for quantitative analysis in this study. A comparison of application writing texts based on the “drive” categorization indicator is shown in Figure 11. The analysis revealed an increase in the frequency of vocabulary related to a sense of community, accomplishment, and honor in composition texts from middle school to college.

Figure 11.

The application text is based on the comparison

Conclusion

The research content of this paper is mainly to propose a method of constructing a sentiment lexicon based on the domain of application writing, and the method has been used to conduct sentiment mining experiments on application writing texts from learners of five types of institutions, and through the experimental research this paper draws the following conclusions:

By analyzing the words in the sample set of recall rate that the dictionary constructed in this paper contains but the other two general dictionaries do not contain, it is found that the dictionary better contains the vocabulary featured by application essay writing. At the same time, the lexicon is able to better discover new words, such as “unknown”, “powerful” and so on. With the large coverage capacity and lexical diversity of the domain sentiment lexicon in this paper, the precision rate is the same as that of the general lexicon, and the recall rate is greatly improved, which indicates that the sentiment lexicon in this paper has strong domain-specific power.

The writing texts from the five types of institutions showed greater variability in both authenticity and affective tone, with the average for middle school writing texts being much lower than the other four types of institutions in terms of the affective tone of the texts. The writing text scores of key universities, second colleges, vocational colleges and high schools are all well above 75, showing obvious positive emotions. From the experimental results, it can be concluded that the method proposed in this paper realizes the multi-dimensional and quantitative interpretation of learners’ emotional tendency in text writing in different institutions. The validity of the method of this paper is also verified.

Language:
English