Error Analysis and Instructional Strategy Adjustment in a Corpus of English Language Learners
Publicado en línea: 19 mar 2025
Recibido: 27 oct 2024
Aceptado: 21 feb 2025
DOI: https://doi.org/10.2478/amns-2025-0513
Palabras clave
© 2025 Fang Wei, published by Sciendo
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Corpus linguistics is the use of computers to analyze and compile large amounts of naturally occurring language [1]. Over the past decades, corpus linguistics has underpinned empirical studies of language change and use, through which more generalized and valid linguistic research conclusions have been drawn. These studies have been categorized into two main types, corpus-based and corpus-driven. Corpus-based research refers to the use of information from a corpus to validate, exemplify, and illustrate known theories and to establish links between data and known theories [2–4]. Corpus-driven research, on the other hand, uses the corpus as a basis to summarize the repetition patterns and frequency distributions in the corpus and to summarize the relevant linguistic theories [5–7]. These two approaches have also become the main corpus research directions at present.
In the process of language learning, word collocation plays an indispensable role. However, the differences between Chinese and English languages and thinking, and improper vocabulary teaching methods have led to the lack of English learners’ word collocation ability [8–9]. Therefore, cultivating students’ awareness of word collocation in teaching and summarizing the collocational behaviors of words with authentic and reliable corpus can help students improve their word collocation ability [10–12]. The corpus allows language teachers and language learners to search and query consecutive texts of tens of millions of words, and these texts are real, living linguistic evidence [13–15]. In each line of the word index of the corpus, the keywords appear centrally, and to the left and right are the words that make up their contexts, and observing and analyzing these words reveals the collocational behavior of the keywords [16–17]. Using the retrieved evidence of the word indexes, it is possible to analyze the collocation behavior of the keywords for error checking, on the basis of which students are able to acquire more comprehensible language [18–19].
In this paper, the classical BP backpropagation algorithm is used to train the recurrent neural network, and in order to solve the problem of long distance dependence of the recurrent neural network, an RNN variant modeling the long and short-term memory network with a gating structure is proposed. Based on the two structures of decoder and encoder, the Seq2Seq framework is established, and the framework is applied to the modeling of text generation tasks. A corpus of learners at home and abroad is listed and categorized, and the English grammar error correction model based on seq2seq is established according to the definition and evaluation criteria of English grammar error correction. Use Soft Attention in global computing combined with seq2seq to deal with the problem of grammar error correction, introduce BN in the model, normalize the input activation parameters of any neuron in each layer of the network, transform the computer text into vectors, and improve the comprehensibility of the corpus semantics through the feedback filtering mechanism of the n-gram language model. The error effect of corpus error correction is analyzed, and based on the analysis results, the adjustment countermeasures of English grammar teaching are proposed.
Recurrent Neural Networks (RNN) are a class of recurrent neural networks that take sequence data as input, recursively in the evolutionary direction of the sequence, and all nodes are connected in a chain fashion [20]. Due to the recursive nature, recurrent neural networks can represent sequences of arbitrary length as vectors of fixed length while focusing on the structured properties of the input.
RNN networks mainly involve some formulas as follows:
Equation (1) contains the parameter matrices
In the training process of recurrent neural networks, the classical BP backpropagation algorithm is usually used to update the parameters forward from the last moment
The biggest difference with the RNN hidden layer unit is that there is an extra memory unit
In the formula (3) of the forgetting gate, the output
The input gates are computed as follows, the tanh function in Eq. (4) produces a new candidate vector
The final output gate
LSTM solves the RNN long-range dependency problem well with three gating structures and achieves better results. However, the complexity of the model is also significantly increased, the model has more parameters, more training data is needed to make the model fit, and the training time of the model is also increased a lot more than RNN.
The GRU unit contains the following main calculation steps:
Firstly, Equation (9) indicates that the hidden state
Sequence to Sequence (Seq2Seq) framework is commonly used for modeling text generation tasks and has achieved great success in areas such as Machine Translation, Speech Recognition, Text Summarization, Question and Answer systems, etc. [21]. The Seq2Seq framework usually consists of two constructs, an Encoder and a Decoder, whose inputs and outputs are both a sequence. In Encoder, the input sequence is usually converted into a fixed-length intermediate vector, and then the intermediate vector is converted into a result sequence by Decoder. A brief structure diagram of the Seq2Seq framework is shown below in Figure 1. The most important feature of this framework is that the output sequence is of variable length, e.g., the input sequence {

Seq2Seq framework structure
Grammatical error correction task can also be regarded as a generative task, the input of the encoder side is the sentence containing grammatical errors, the output of the decoder side is the sentence predicted to be generated by the model after correction, at the same time in the correction of the error may be inserted or deleted some words resulting in the original sentence and the target sentence length inconsistency. For example, the sentence "I always study English in morning." The article "the" is missing from "in morning", and the Seq2Seq framework can be added by decoding the word "in" at the next decoded position when decoding to the word "in". This method of predicting the generation of indefinite-length sentences can achieve the correction of grammatical errors contained in the original sentence in the process of generating sentences.
NUCLE
NUCLE is a learner corpus co-constructed by the Natural Language Processing team at NUS. The corpus contains 1,397 English essays written by NUS undergraduates on topics in the fields of science and technology, healthcare, and economics, with more than one million words. All compositions are manually marked for grammatical errors by professional English teachers.
FCE
FCE is a public subset of the Cambridge Learner Corpus, which consists of 1,244 English test responses written by ESL learners in the form of short essays, letters, and descriptions. All texts are manually annotated with grammatical errors by an annotator, and all grammatical errors are categorized by category.The FCE corpus is also a publicly available common training corpus for grammatical error correction tasks, and the corpus itself is divided into training, development, and test sets, although when in use, the FCE corpus is usually used only in the training phase.
Lang-8
The Lang-8 corpus of English language learners is a subset of the multilingual Lang-8 corpus extracted and preprocessed in 2012. The corpus contains 100,000 English texts written by ESL learners on social learning sites, and the corpus is also more noisy because the texts are annotated for grammatical errors by native English speakers on forums rather than by professional linguistic corpus annotators. Nevertheless, Lang-8 is still by far the largest publicly available learner corpus for the task of grammatical error correction, which contains a total of 1.04M parallel sentence pairs and about 11.86M Token counts, while the corpus quality is higher than the majority of synthetic corpora.
JFLEG
Unlike NUCLE and other corpora, the annotation of JFLEG corpus includes not only the correction of sentence grammatical errors, but also sentence rewriting at the fluency level, which requires the grammar error correction system to not only correct the grammatical errors, but also to make the corrected sentences have a more fluent expression.
There are fewer publicly available manually annotated corpora for grammar error correction tasks in China, and this subsection focuses on CLEC, which is used in this paper.
The CLEC corpus collects English writing compositions from students with different English proficiency levels, including secondary school students, university English level 4 and 6, and professional English level 4 and 6, corresponding to the five categories from ST2 to ST6, and containing more than one million words in total. The purpose of this corpus is to observe and analyze Chinese students’ verbal errors in the use of English through quantitative and qualitative methods, so as to provide useful information for the development of English education in China. For the grammar correction task, the grammatical errors in the CLEC corpus reflect the English grammar usage habits of Chinese students, and the use of the CLEC corpus to train and test the grammar correction model allows the grammar correction model to be better applied to the grammar correction scenarios of Chinese English learners.
CoNLL2013 and CoNLL2014 are two competitions held specifically for the problem of English grammatical error correction, which gave a clearer definition of the problem of English grammatical error correction and attracted many teams in the field of grammatical error correction to participate. It also promoted the enthusiasm of academics in researching this problem, which led to a certain development of grammar error correction.The task given in CoNLL2013 was to correct five common grammatical errors in English, including prepositional errors, noun singular-plural errors, qualifier errors, subject-verb agreement errors, and verb morphology errors. In CoNLL2014, on the other hand, the 28 most common grammatical errors in English are corrected, which is more difficult than CoNLL2013 but also more practical to the problem of English grammar correction. In this paper, the problems in CoNLL2014 will be studied, and natural language processing techniques and methods will be used to improve the effect of grammatical error correction.
The open source tool MaxMatchscorer implements the MaxMatch algorithm, and gives a simple and easy to use, you can directly use the tool to evaluate the effect of model error correction. The following is an introduction to the principle of MaxMatch algorithm.
Correction rate
Correction rate
The numerator of both equations above is the number of sentences in which the model correction result matches the reference answer, the denominator of the correction rate
From the above equation, it can be seen that the algorithm amplifies the percentage contribution of the correction rate and reduces the percentage contribution of the correction rate. It makes the overall presentation that whenever a sentence is corrected it is better to be correct, otherwise it is preferred not to make corrections, penalizing the phenomenon of correcting an otherwise correct sentence.
Aiming at the fact that the English grammar error correction problem has certain similarities with machine translation, and the seq2Seq model has made many breakthroughs in the field of NLP, in this paper, we will choose seq2Seq as the basic model for grammar error correction problem solving, and improve the training efficiency and model error correction ability of the basic model of seq2Seq, so as to improve the effect of English grammar error correction. The main improvement points are described in detail below.
In the seq2seq model, the encoder forms the input sequence into a fixed-length context vector, and a large amount of detailed information is lost during the computation of the context vector, which becomes more obvious when the sentence length exceeds a certain limit. Therefore, scholars have carried out a series of studies to address this problem, which was eventually inspired by research work on human vision.
Figure 2 shows the attention model, according to the calculation range division of the input sequence, the attention mechanism can be divided into two kinds of local and global, in this paper, we will use Soft Attention in the global computation combined with seq2seq to deal with the problem of grammatical error correction, so the following will be introduced to Soft Attention [22].

Soft Attention model
The value of weight
Deep neural networks are powerful, but their requirements for machine computational power are high, and the improvement of training efficiency plays a key role in model optimization and improvement [23]. Theory shows that normalizing the input activation parameters of neurons can reduce the model training time. The (BN) introduced in CNN by normalizing the input activation parameters of any neuron in each layer of the network, the activation parameters conform to the standard normal distribution after the transformation, which makes the activation parameters fall in the region of the slope of the nonlinear function, the gradient is larger, and the model convergence speed is faster.
However, since the BN operation depends on the mini-batch first-order and second-order statistics, which are related to the size of the mini-batch, and the RNN input length is variable, the BN cannot be directly used in the RNN. In order to improve the RNN training efficiency, researchers proposed (LN), which can significantly reduce the model training time and various RNN networks after layer normalization outperform the original network.
LN is normalized for the training samples, which is performed independently and in parallel with other sample data, so it is not as demanding for the data distribution as BN, and the training difficulty decreases drastically.LN is a kind of transversal normalization, which maps the inputs of neurons of an entire layer in an RNN into the same distribution by summing up the inputs of neurons in the layer with the computation of the variance, and all the hidden elements share the same normalization terms
Since computers can not understand natural language, so the use of computers for text processing must be vectorized into numbers that computers can understand. The simplest way to encode is one-hot matrix representation, that is, with a fixed length of the size of the vector represents a word, the vector length for the size of the word list, only in the word appears in the position of 1, and the other are 0, such as “book” using [0, 0, 1, 0, 0, 0, …] that the method is simple, but the disadvantages are obvious. The method is simple, but the disadvantages are also obvious, the data dimension is too large, too sparse, and the relationship between words is difficult to reflect.
For the modification suggestions submitted by users, taking into account the varying levels of English proficiency of users, it is necessary to screen these utterance suggestions to identify which are correctly modified utterances and which are incorrect utterances. This is related to the process of re-learning and re-training of the system, and belongs to the key module of the system. This filtering process is actually comparing the modified statements given by the system and the modified suggested statements given by the user which one is more credible, then the sentences need to be scored, and those with higher scores can be considered more probable to be the sentences with no grammatical errors, so when the user suggests to modify the sentences with a score lower than the system’s score of the modified sentences will not be adopted. Therefore, in order to effectively filter invalid suggested text, the n-gram grammar model is used to score the sentences. The n-gram model is described below.
Using a string S to represent a sentence, the n-gram model models the string S. P(S) denotes the probability of occurrence of the sentence. Words are the basic units in the n-gram, and a sentence consisting of m words has probability
According to the above equation, it can be seen that the
For this reason, the Markov assumption is introduced that the current word is only related to the previous few words, which can simplify the calculation process as:
For the n-gram language model, it is common that n can be taken as 1, 2, or 3. A start and end flag is usually added before and at the end of the sentence to facilitate uniform computation.
When n=2, bigram is:
When n = 3, trigram is:
It is sufficient to make Eq. (23) obtain the maximum value according to the maximum likelihood estimation method. These are the basic elements of the N-metric grammar. In the training corpus, the value of
After training the model it is possible to calculate the probability and perplexity of the sentence, and the concepts of probability and perplexity are first briefly explained below.
Probability, as the name suggests, is the frequency of occurrence of the sentence, according to the basics of the N-meta grammatical model introduced earlier:
Perplexity, the basic idea is to give normal sentences to the larger the probability value of the language model is better, after training the model, use the model to score the test set, at this time, the test set are all normal sentences, then the larger the score of the model is better. The relevant formula is as follows:
It can be rewritten according to the chain rule:
From the formula, it can be seen that the smaller the confusion degree is, the larger the sentence probability is. Comparing probability and perplexity, it can be found that the calculation of probability is greatly influenced by the length of the sentence and the number of words, while perplexity can effectively reduce this influence, so perplexity is often used instead of probability to judge the sentence and model. Therefore, in this paper, the confusion degree is also used, but for the sake of convenience and semantic comprehensibility, probability is sometimes used instead.
In order to demonstrate the effectiveness of the attention mechanism proposed in this paper, this subsection compares the error correction effects of the model without the attention mechanism with the present model with the attention mechanism, as shown in Fig. 3. The values in the figure are the P, R, F0.5 results obtained on the CoNLL-2014 test set. It can be seen that the model with the addition of the corresponding attention mechanism is significantly better than the model without attention on all five types. Among them, the attention mechanism is the most helpful for the noun singular-plural model, which is 25 more effective than the one without the attention mechanism on the F0.5 metric.

The attention mechanism is relatively effective
This subsection compares the error correction results of the three classification models, this paper’s model, the CUUI model, and the deep contextual model, on five error types, namely, articles, prepositions, verb morphology, noun singular-plural, and subject-verb agreement, as shown in Figure 4. It can be seen that this paper’s model achieves the highest results in the other 4 types except for the preposition type, and in the F0.5 metric, the difference between the classification results of this paper’s model for coronal, verb morphology, noun singular and plural, and subject-predicate agreement is 6.6, 18.1, 8.0, and 8.0, when compared to the deep contextual model. And the progress is relatively obvious. In terms of preposition types, the model in this paper is slightly inferior to the deep contextual model, probably due to the large number of preposition classes, the deep model still can’t get a better grasp of the collocation laws.

Comparison of results based on classification
Two classes were selected as experimental subjects in college J. The experimental class used the corpus error analysis constructed in this paper for English language learning, while the control panel used the traditional English language learning method for 15 weeks.
A pre-study test was administered to these two classes when they first entered their junior year of college, and the results showed that students in both classes had more grammatical errors in English writing, concentrating on prepositions, verbs, and conjunctions, and both had many common problems. After entering the third year of college, students were more concerned about improving their English performance, and after two years of study, students’ reading and completing the blanks were more correct, which was greatly related to the fact that English teachers in the school emphasized reading. At the stage of freshman and sophomore years, students practiced reading and gap-filling exercises a lot, but less practice in written expression and even less correction of errors, and usually memorized model essays and wrote essays for exams, so there are more grammatical errors in students’ writing, and the present study carries out an action research on the treatment of grammatical errors in high school students’ English writing according to the problem of grammatical errors in students’ writing.
Before, after the first, second and third rounds of this action research, tests will be given to the students of both classes, and then the test scores of both classes will be analyzed according to the data analysis method, so as to judge the effectiveness of the grammar correction material bank constructed in this paper on the students’ written expressions in English.
By correcting the pre-test paper, the results of the pre-test paper according to the eight lexical aspects examined in grammar [prepositional errors, pronoun errors, coronal errors, noun errors, adjective errors, adverb errors, verb errors and conjunction errors (syntactic errors)] are shown in Table 1.
In general, it seems that the students’ error rate in pronouns is low, but there is a compound structure, the students will be confused by other grammatical points, ignoring the use of pronouns in the wrong way, in topic 13, 78.15% of the students revised correctly, but in the other topic (topic 11), the correct rate is only 34.79%, and most of them directly ignored not to revise the error here and did not understand the grammatical examination of this sentence.
Test volume score
Subject number | Exploratory morphology | Scoring rate | Subject number | Exploratory morphology | Scoring rate |
---|---|---|---|---|---|
1 | Nouns | 82.65% | |||
2 | Article | 82.65% | 12 | Verbs | 69.36% |
3 | Article | 73.45% | Pronoun | ||
4 | Nouns | 26.06% | 14 | Verbs | 21.65% |
5 | Adverb | 85.66% | 15 | Verbs | 69.52% |
6 | Adjective | 65.18% | 16 | Conjunction | 78.23% |
7 | Article | 43.66% | 17 | Conjunction | 78.23% |
8 | Preposition | 56.96% | 18 | Conjunction | 56.58% |
9 | Preposition | 60.83% | 19 | Verbs | 69.15% |
10 | Verbs | 60.84% | 20 | Conjunction | 4.36% |
Table 2 shows the statistical scale of the sample mean of the writing pre-test, and the score rate of the essay fill-in-the-blank is higher, which is 12.9352. During the period of freshman and sophomore years, the students practiced a lot of grammar fill-in-the-blank by using the English grammar correction model in the corpus, and they mastered this type of question well, and the prompt words of the essay fill-in-the-blank questions are more obvious, which makes it easier for the students to write the correct answers compared with short-text corrections.
The sample mean statistics before writing
Type | Mean value | Case number | Standard deviation | Standard error mean |
---|---|---|---|---|
Simple Sentence | 12.1654 | 50 | 2.7563 | 0.3648 |
Error correction | 5.7556 | 50 | 2.6466 | 0.3896 |
Single Fill | 13.6485 | 50 | 2.5489 | 0.3248 |
Single Sentence Translation | 5.9547 | 50 | 2.9954 | 0.3495 |
Fill In | 50 | 2.3964 | 0.3685 | |
Written Expression | 14.9487 | 50 | 3.2874 | 0.3185 |
Table 3 shows the results of the pre- and post-tests of the first round of action research. Before and after the first round of action research, the score rate of the five types of questions, such as prepositional, pronouns, nouns, predicate verbs, and parallel conjunctions, increased, but the score rate of the five types of questions, such as coronal, adjectives, adverbs, nonpredicate verbs, and subordinate conjunctions, on the contrary, declined. The type with the highest increase in score rate was the noun category, which reached 29.89%, and the type with the highest decrease was the adjective category questions, which reached -31.46%. Taken together, before and after the first round of action research, although the students’ individual category score fluctuation values varied greatly, the final average score had little ups and downs and was more influenced by randomness.
The first round of action was studied
/ | The first round of action research (score rate%) | The first round of action (score rate%) | Fluctuating value |
---|---|---|---|
Preposition | 58.64% | 66.34% | 7.70% |
Article | 66.24% | 58.71% | -7.53% |
Pronoun | 56.48% | 82.63% | 26.15% |
Adjective | 65.21% | 33.75% | -31.46% |
Adverb | 86.69% | 71.69% | -15.00% |
Nouns | 54.36% | 84.25% | |
Predicate verb | 53.21% | 57.63% | 4.42% |
Non-predicate verb | 65.97% | 53.23% | -12.74% |
Coordinate conjunction | 67.15% | 70.64% | 3.49% |
Subordinate conjunction | 41.36% | 41.26% | -0.10% |
Average score | 6.18 | 6.23 | 0.05 |
Pair difference | |||
Average (post-test) | Standard deviation | Standard error mean | The difference is 95% |
0.48% | 18.654% | 5.8965% | -12.856% |
Confidence interval limit | T | Freedom | Significance (double tail) |
13.851% | 0.0815 | 10 | 0.945 |
A paired sample t-test was conducted on the grades before and after the first round of action research. The value of significance was 0.945, indicating that there was no significant correlation between the changes in pre- and post-test scores and the first round of action research. Combined with the fluctuations in the score rates of the various types of grammar questions, it can be concluded that the first round of action research using the corpus for error analysis in EFL did not have a significant impact on the students’ knowledge of grammar.
Table 4 shows the results of the pre-test and post-test of the second round of action research. Before and after conducting the second round of action research, the score rate of the grammar topics in all categories increased, except for the adverbial category, where the score rate decreased slightly. Among them, the type with the highest increase in the score rate is the predicate verb category, which reaches 32.67%. Taken together, the students’ grades in the grammar questions before and after the second round of action research on English language learning using the error analysis in the corpus constructed in this paper have increased significantly. The fluctuation value of the average score also rose more significantly compared to the first round of action research.
The first test of the second round of action and the subsequent test results
/ | The first test of the second round of action (score rate%) | After the second round of action (score rate) | Fluctuating value |
---|---|---|---|
Preposition | 66.45% | 76.09% | 9.64% |
Article | 58.73% | 86.48% | 27.75% |
Pronoun | 82.63% | 88.63% | 6.00% |
Adjective | 33.54% | 63.15% | 29.61% |
Adverb | 71.85% | 65.05% | -6.80% |
Nouns | 84.59% | 92.38% | 7.79% |
Predicate verb | 57.56% | 90.23% | 32.67% |
Non-predicate verb | 53.69% | 75.63% | 21.94% |
Coordinate conjunction | 70.65% | 71.26% | 0.61% |
Subordinate conjunction | 41.65% | 43.42% | 1.77% |
Average score | 6.15 | 6.69 | 0.54 |
Pair difference | |||
Average (post-test) | Standard deviation | Standard error mean | The difference is 95% |
13.10% | 13.76% | 4.43% | 3.36% |
Confidence interval limit | T | Freedom | Significance (double tail) |
22.92% | 3.039 | 9 | 0.0154 |
A paired samples t-test was conducted on the scores before and after the second round of action research. This time, the t-value was 3.039 and the significance value was 0.0154, indicating that the changes in the pre- and post-test scores were significant and statistically significant. Combined with the fluctuations in the score rates of the various types of grammar questions, it can be concluded that the second round of the action research significantly improved the students’ mastery of the various types of grammar.
Table 5 shows the statistics comparing the pre and post-test scoring rates for the third round of actions, with significant increases in the pre and post-test scoring rates for Topics 1 and 4 showing a significant improvement in the students’ understanding of the use of nouns. Topics 2, 3 and 7 showed an increase in score rate of 9.11%, a decline of -8.40% and an increase of 25.50%, respectively, which indicates that students’ understanding of coronal use is not always consistent and may need to be strengthened in some areas. Adverbs were examined in Topic 5 and the score increased from 86.93% to 100%, showing that students have a basic grasp of the use of adverbs. Adjectives were examined in Topic 6, and there was a 21.03% increase in the score, showing some improvement in the mastery of adjectives. Prepositions were examined in Topics 8 and 9 with an increase in score of 13.31% and 30.42% respectively, showing a significant improvement in students’ understanding of the use of prepositions. Verbs appeared in Topics 10, 12, 14, 15 and 19 with a range of improvement in scores from 17.1% to 25.36%, which shows some improvement in students’ understanding of verb usage. Pronouns were examined in Topics 11 and 13 and there was an increase in score of 34.49% and 17.28%, which shows a greater improvement in students’ use of pronouns. Conjunctions were examined in Topics 16, 17, 18, and 20, and there was an increase of more than 10% in the score rate, which shows a greater improvement in students’ understanding of the use of conjunctions.
The third round is compared to the rate of score
Subject number | Exploratory morphology | Pretest rate | After score | Rate fluctuation |
---|---|---|---|---|
1 | Nouns | 82.63% | 100% | 17.37% |
2 | Article | 82.14% | 91.25% | 9.11% |
3 | Article | 73.66% | 65.26% | -8.40% |
4 | Nouns | 26.06% | 60.25% | 34.19% |
5 | Adverb | 86.93% | 100% | 13.07% |
6 | Adjective | 65.23% | 86.26% | 21.03% |
7 | Article | 43.98% | 69.48% | 25.50% |
8 | Preposition | 56.31% | 69.62% | 13.31% |
9 | Preposition | 60.84% | 91.26% | 30.42% |
10 | Verbs | 60.89% | 86.25% | 25.36% |
11 | Pronoun | 34.66% | 69.15% | 34.49% |
12 | Verbs | 69.36% | 91.62% | 22.26% |
13 | Pronoun | 78.15% | 95.43% | 17.28% |
14 | Verbs | 21.96% | 39.65% | 17.69% |
15 | Verbs | 69.26% | 86.36% | 17.10% |
16 | Conjunction | 78.56% | 91.26% | 12.70% |
17 | Conjunction | 78.26% | 100% | 21.74% |
18 | Conjunction | 56.23% | 78.26% | 22.03% |
19 | Verbs | 69.48% | 86.64% | 17.16% |
20 | Conjunction | 4.36% | 26.97% | 22.61% |
Taken together, the majority of students improved their understanding of all types of lexemes. Although there was a decrease in scores on one quiz for articles, there was also an increase in scores on other quizzes, and it is possible that this is an isolated phenomenon. These data show improvement in students’ grammatical understanding and application.
Table 6 shows the paired samples t-test for the comparison analysis of the pre-test and the writing post-test, which showed that the post-test, compared to the pre-test, improved by an average of 5.0645 points for single-sentence corrections, 3.0354 points for short-text corrections, 3.3214 points for one-sentence fills in the blanks, 1.7258 points for one-sentence translations, 3.3454 points for essay fills in the blanks and 1.4598 points. Although the increase in the scores of each question type is different, it proves that students have made some progress.
Test of matching sample t after test and writing
Item | Mean value | Standard deviation | |
---|---|---|---|
Simple sentence | Post test-Pre test | 5.0645 | 3.4215 |
Error correction | 3.0354 | 4.0695 | |
Single fill | 3.3214 | 3.8545 | |
Single sentence translation | 1.7258 | 3.1698 | |
Fill in | 3.3454 | 4.6284 | |
Written expression | 1.4598 | 2.3654 | |
Pairing difference | |||
Item | Mean standard error | Difference 95% confidence interval | |
Lower limit | Upper limit | ||
Simple sentence | 0.5261 | 4.0069 | 6.0684 |
Error correction | 0.6598 | 1.8935 | 4.2254 |
Single fill | 0.5621 | 2.1658 | 4.4825 |
Single sentence translation | 0.4628 | 0.7985 | 2.6365 |
Fill in | 0.9156 | 1.4585 | 4.2158 |
Written expression | 0.3515 | 0.7856 | 2.2365 |
Item | T | Freedom | Significance (double tail) |
Simple sentence | 9.7184 | 45 | 0.000 |
Error correction | 5.1254 | 45 | 0.000 |
Single fill | 5.8596 | 45 | 0.000 |
Single sentence translation | 3.7185 | 45 | 0.000 |
Fill in | 3.3985 | 45 | 0.000 |
Written expression | 4.2615 | 45 | 0.000 |
In summary, the post-test performed better than the pre-test in all of the items mentioned. The t-value proves that such an improvement is statistically significant and the significance value of 0 for each item also indicates the significance of the improvement. This means that all the instructional activities, be it single sentence correction, short correction, single sentence fill in the blank, single sentence translation, essay fill in the blank, or written expression have achieved effective improvement.
Figure 5 shows the ontology errors of the experimental class control class before and after the test, as can be seen from the figure: Regarding case errors, before the experiment, the difference between the experimental class and the control class in terms of case errors is not very large, but after the experiment, the control class kept the same as before the experiment which is basically stable, but the experimental class basically did not make any mistakes in terms of case after the experiment. Regarding punctuation, before the experiment, the number of errors in the experimental class and the control class were 16 and 18 respectively, which is not a big difference. But after the experiment, the experimental class made good results in punctuation errors, 4 compared to the previous 16 errors, which proves that the teaching about punctuation is basically successful in the process of experimental teaching based on corpus-based error analysis. Regarding word spelling errors, before the experiment, the difference between the experimental class and the control class is not very much, but after the experiment, the data about the control class has increased, while the experimental class and the previous error rate compared to reduce by half, is not a very obvious progress, but words are after all, a process of progress over time, so to a certain extent, this aspect still attracts the attention of students.

The experimental class is wrong after the control of the class
To sum up, the teaching strategies and methods of corpus error analysis applied in the experiment are helpful to students regarding ontological error correction.
According to the survey of students’ English grammar test scores, it can be seen that students’ poor knowledge of grammar and vague concepts of grammar lead to a large number of errors in oral or written expression in English.
Language knowledge and language skills are the basis of comprehensive language use ability, and grammar is an important part of language knowledge, so to cultivate students’ comprehensive language use ability, it is necessary for students to master a certain amount of grammar knowledge. Without a solid foundation of grammatical rules, it is impossible to learn any foreign language, and language learning must follow the law of language learning. The acquisition of language without explanation and grammar refers to the natural acquisition of the mother tongue by young children, while today’s foreign languages must be acquired through classroom learning due to the lack of environmental support. Therefore, teachers should pay attention to the teaching of grammar, familiarize themselves with the syllabus, teach and summarize the basic grammatical knowledge when teaching English, and strengthen the grammatical rules by carrying out a lot of training in listening, speaking, reading and writing, so that they can familiarize themselves with the rules of English grammar and encourage students to communicate with each other by applying the forms of the language they have learned, organically combining the forms of the language with the functions of communication in a certain scenario, and improving the ability of comprehensive use of the language.
In a word, the cultivation of comprehensive language ability should be emphasized in English teaching, but the teaching of language knowledge should not be taken lightly, and attention should be paid to the cultivation of learners’ grammatical awareness.
Grammar has to be taught, and it has to be taught well. To teach grammar well, method is the key. This section discusses the methods of teaching English grammar in secondary schools in view of the problems in grammar learning of secondary school students. This section will mainly discuss the use of comparative method to teach grammar, the combination of context to teach grammar and the combination of task-based teaching method to teach grammar.
Due to the differences between English and Chinese languages, students often make mistakes when using English because of the influence of mother tongue transfer, therefore, it is quite important to teach English grammar with appropriate use of comparative method so that students can understand the differences between English and Chinese languages.
In the process of grammar teaching, if only through the words, phrases, sentences and other static forms of teaching, repeated drills to make students memorize the grammar rules, this will only make students learn boring, and can not really understand the rules of grammar, can not flexibly use the language. Combined with the context of teaching grammar, that is, in teaching according to the specific teaching content and characteristics of the design of different contexts, the abstract rules of grammar into concrete language facts, in the corresponding context for students to actively feel and discover the rules of grammar, so that students will really learn and remember the rules of grammar, to avoid inappropriate generalization of the rules of grammar, to reduce the grammatical errors within the language.
The task-based teaching method emerged in the century. Task-based language teaching advocates letting students do things, i.e., mastering language through accomplishing various tasks, and adopting the method of letting students discover, generalize and summarize to cultivate their ability of independent learning and inquiry. The core idea of task-based teaching is to simulate the kinds of activities that people engage in when they use language in school life and society, to combine language teaching with language use, and to let learners actively participate in the attempts to communicate in the target language, so as to cultivate students’ comprehensive language competence. Advocates of task-based language teaching believe that grammar teaching should not be rejected, but should be emphasized instead. They believe that classroom instruction should teach the forms of grammar and how these forms can be used for communicative purposes.
It is one of the important tasks of the English language program for teachers to consciously strengthen the guidance of learning strategies for students so that they can develop good learning habits. Teachers should help students gradually learn how to learn English and how to master English grammar concepts in the process of learning and using the English language, and they should pay attention to cultivating students’ habit of thinking in English. Teachers can strengthen the explanation and training of grammar knowledge and teach students to master and use grammar knowledge skillfully. Major language phenomena such as key sentence patterns, tenses and morphology should be repeatedly practiced, which will constantly reinforce the stimulation of students’ brains and is an effective means to overcome the interference of Chinese sentence patterns. Teachers can pay attention to the combination of the usual reading class and writing training, so that students can extract the typical sentence patterns learned in the reading class and drill them repeatedly in order to master them. Students can be introduced to some of the sentence transformation method at the right time, such as a variety of subordinate clauses, a variety of connectives, a variety of expressions used to emphasize the stressed sentence, exclamatory sentence, inverted sentence, etc., the usual sentence pattern. Students can master various grammars and syntaxes and make fewer mistakes.
Teaching grammar in the right way will enable students to master grammar knowledge better and reduce grammatical errors in language use. However, the process of language learning is a process of traveling with mistakes, and it is inevitable for students to make grammatical mistakes. Teachers should deal with students’ grammatical mistakes correctly in grammar teaching, so that students can learn from their mistakes and improve their grammatical level and language ability.
In this paper, we use deep neural network, based on foreign and domestic learner corpus, to establish English grammar error correction model based on seq2seq, add the attention mechanism, process the grammar error correction problem, introduce layer normalization, and perform text processing before vectorization operation. In order to filter the invalid suggestion text effectively, n-gram grammar model is used to score the sentences and complete the feedback filtering of English semantics. The model of this paper has a classification result difference of 6.631, 18.107, 7.944, and 8.007 for coronal, verb morphology, noun singular-plural, and subject-predicate agreement compared with the deep context model in the F0.5 metrics.In the pre-test of grammatical error correction, 78.15% of the students modified the use of pronouns correctly, but the correct rate was only 34.79% in the other question. Most of them directly ignored not modifying the error here and did not understand the grammatical examination point of this sentence. In the third round of the action post-test, the score rate of each topic has been improved to different degrees, the examination of adverbs in topic 5, the score rate has been increased from 86.93% to 100%, and the students have basically mastered the use of adverbs. Pronouns were examined in Topics 11 and 13, and the score rate has increased from 34.49% and 17.28%, which shows that students have improved their use of pronouns. By comparing the experiments and analyzing the ontology errors before and after the error analysis of the corpus, the number of errors in the experimental class and the control class before the experiment were 16 and 18 respectively, and the experimental class after the experiment made 4 errors in punctuation, which is a good achievement compared with the previous one, which proves that, in the process of the experimental teaching based on the error analysis of the corpus, the teaching about the punctuation in the English grammar is basically successful.