A Comparative Evaluation of Deep and Shallow Approaches to Common Grammatical Errors
by user
Comments
Transcript
A Comparative Evaluation of Deep and Shallow Approaches to Common Grammatical Errors
A Comparative Evaluation of Deep and Shallow Approaches to the Automatic Detection of Common Grammatical Errors Joachim Wagner, Jennifer Foster, and Josef van Genabith EMNLP-CoNLL 28th June 2007 National Centre for Language Technology School of Computing, Dublin City University 1 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 2 Why Judge the Grammaticality? • Grammar checking • Computer-assisted language learning – Feedback – Writing aid – Automatic essay grading • Re-rank computer-generated output – Machine translation 3 Why this Evaluation? • No agreed standard • Differences in – – – – What is evaluated Corpora Error density Error types 4 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 5 Deep Approaches • Precision grammar • Aim to distinguish grammatical sentences from ungrammatical sentences • Grammar engineers – Avoid overgeneration – Increase coverage • For English: – ParGram / XLE (LFG) – English Resource Grammar / LKB (HPSG) 6 Shallow Approaches • Real-word spelling errors – vs grammar errors in general • Part-of-speech (POS) n-grams – – – – – Raw frequency Machine learning-based classifier Features of local context Noisy channel model N-gram similarity, POS tag set 7 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 8 Common Grammatical Errors • 20,000 word corpus • Ungrammatical English sentences – Newspapers, academic papers, emails, … • Correction operators – – – – Substitute (48 %) Insert (24 %) Delete (17 %) Combination (11 %) 9 Common Grammatical Errors • 20,000 word corpus • Ungrammatical English sentences – Newspapers, academic papers, emails, … • Correction operators – – – – Substitute (48 %) Insert (24 %) Delete (17 %) Combination (11 %) Agreement errors Real-word spelling errors 10 Chosen Error Types Agreement: She steered Melissa around a corners. Real-word: She could no comprehend. Extra word: Was that in the summer in? Missing word: What the subject? 11 Automatic Error Creation Agreement: replace determiner, noun or verb Real-word: replace according to pre-compiled list Extra word: duplicate token or part-of-speech, or insert a random token Missing word: delete token (likelihood based on part-of-speech) 12 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 13 BNC Test Data (1) BNC: 6.4 M sentences 4.2 M sentences (no speech, poems, captions and list items) Randomisation 1 2 3 4 5 … 10 10 sets with 420 K sentences each 14 BNC Test Data (2) Error corpus 1 2 3 4 5 … 10 1 2 3 4 5 1 2 3 4 5 1 2 3 4 5 1 2 3 4 5 … 10 Error creation Agreement Real-word … … 10 10 Extra word Missing word … 10 15 BNC Test Data (3) Mixed error type ¼ each 1 1 1 1 … 1 … … ¼ each 10 10 10 10 10 16 BNC Test Data (4) 5 error types: agreement, real-word, extra word, missing word, mixed errors 1 50 sets 1 … 10 1 1 1 … 10 10 10 1 1 … 10 10 1 … 10 10 1 1 … 10 10 Each 50:50 ungrammatical:grammatical 17 BNC Test Data (5) Test data 1 1 1 1 1 1 1 1 1 1 2 2 2 2 2 2 2 2 2 2 … 10 … 10 10 10 … 10 10 … 10 10 … 10 10 Example: 1st crossvalidation run for agreement errors Training data (if required by method) 18 Evaluation Measures • • • • • tp = true positive tn = true negative fp = false positive fn = false negative Precision tp / (tp + fp) Recall tp / (tp + fn) F-score 2*pr*re / (pr + re) Accuracy (tp + tn) / total tp := ungrammatical sentences identified as such 19 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 20 Overview of Methods XLE Output POS n-gram information M1 M2 Basic methods M3 M4 Decision tree methods M5 21 M1 Method 1: Precision Grammar • XLE English LFG • Fragment rule – Parses ungrammatical input – Marked with * • Zero number of parses • Parser exceptions (time-out, memory) 22 M1 XLE Parsing First 60 K sentences 1 … 10 1 … 10 1 … 10 1 … 10 1 … 10 XLE 50 x 60 K = 3 M parse results 23 M2 Method 2: POS N-grams • Flag rare POS n-grams as errors • Rare: according to reference corpus • Parameters: n and frequency threshold – Tested n = 2, …, 7 on held-out data – Best: n = 5 and frequency threshold 4 24 M2 POS N-gram Information 9 sets 1 Reference n-gram table Repeated for n = 2, 3, …, 7 … 10 1 … 10 1 … 10 1 … 10 1 … 10 Rarest n-gram 3 M frequency values 25 M3 Method 3: Decision Trees on XLE Output • Output statistics – Starredness (0 or 1) and parser exceptions (-1 = time-out, -2 = exceeded memory, …) – Number of optimal parses – Number of unoptimal parses – Duration of parsing – Number of subtrees – Number of words 26 M3 Decision Tree Example Star? >= 0 <0 Star? U <1 >= 1 Optimal? <5 U = ungrammatical G = grammatical U >= 5 U G 27 M4 Method 4: Decision Trees on Ngrams • Frequency of rarest n-gram in sentence • N = 2, …, 7 – feature vector: 6 numbers 28 M4 Decision Tree Example 5-gram? >= 4 <4 7-gram? U <1 >= 1 5-gram? <45 U >= 45 G G 29 M5 Method 5: Decision Trees on Combined Feature Sets Star? >= 0 <0 Star? U <1 >= 1 5-gram? <4 U >= 4 U G 30 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 31 F-Score Strengths of each Method 0.8 0.7 0.6 0.5 Agreement Real-word Extra word Missing word Mixed errors 0.4 0.3 0.2 0.1 0.0 XLE ngram XLE+DT ngram+DT combined 32 F-Score Comparison of Methods 0.8 0.7 0.6 0.5 XLE ngram XLE+DT ngram+DT combined 0.4 0.3 0.2 0.1 0.0 Agreement Real-word Extra word Missing word Mixed errors 33 Results: F-Score 0.8 0.7 0.6 0.5 XLE ngram XLE+DT ngram+DT combined 0.4 0.3 0.2 0.1 0.0 Agreement Real-word Extra word Missing word Mixed errors 34 Talk Outline • • • • • • • Motivation Background Artificial Error Corpus Evaluation Procedure Error Detection Methods Results and Analysis Conclusion and Future Work 35 Conclusions • Basic methods surprisingly close to each other • Decision tree effective with deep approach • Combined approach best on all but one error type 36 Future Work • Error types: – Word order – Multiple errors per sentence • • • • Add more features Other languages Test on MT output Establish upper bound 37 Thank You! Djamé Seddah (La Sorbonne University) National Centre for Language Technology School of Computing, Dublin City University 38 Extra Slides • • • • • • • P/R/F/A graphs More on why judge grammaticality Precision Grammars in CALL Error creation examples Variance in cross-validation runs Precision over recall graphs (M3) More future work 39 Results: Precision 0.8 0.7 0.6 0.5 XLE ngram XLE+DT ngram+DT combined 0.4 0.3 0.2 0.1 0.0 Agreement Real-word Extra word Missing word Mixed errors 40 Results: Recall 0.8 0.7 0.6 0.5 XLE ngram XLE+DT ngram+DT combined 0.4 0.3 0.2 0.1 0.0 Agreement Real-word Extra word Missing word Mixed errors 41 Results: F-Score 0.8 0.7 0.6 0.5 XLE ngram XLE+DT ngram+DT combined 0.4 0.3 0.2 0.1 0.0 Agreement Real-word Extra word Missing word Mixed errors 42 Results: Accuracy 0.8 0.7 0.6 0.5 XLE ngram XLE+DT ngram+DT combined 0.4 0.3 0.2 0.1 0.0 Agreement Real-word Extra word Missing word Mixed errors 43 Results: Precision 0.8 0.7 0.6 0.5 Agreement Real-word Extra word Missing word Mixed errors 0.4 0.3 0.2 0.1 0.0 XLE ngram XLE+DT ngram+DT combined 44 Results: Recall 0.8 0.7 0.6 0.5 Agreement Real-word Extra word Missing word Mixed errors 0.4 0.3 0.2 0.1 0.0 XLE ngram XLE+DT ngram+DT combined 45 Results: F-Score 0.8 0.7 0.6 0.5 Agreement Real-word Extra word Missing word Mixed errors 0.4 0.3 0.2 0.1 0.0 XLE ngram XLE+DT ngram+DT combined 46 Results: Accuracy 0.8 0.7 0.6 0.5 Agreement Real-word Extra word Missing word Mixed errors 0.4 0.3 0.2 0.1 0.0 XLE ngram XLE+DT ngram+DT combined 47 Why Judge Grammaticality? (2) • Automatic essay grading • Trigger deep error analysis – Increase speed – Reduce overflagging • Most approaches easily extend to – Locating errors – Classifying errors 48 Precision Grammars in CALL • Focus: – Locate and categorise errors • Approaches: – Extend existing grammars – Write new grammars 49 Grammar Checker Research • Focus of grammar checker research – – – – Locate errors Categorise errors Propose corrections Other feedback (CALL) 50 N-gram Methods • Flag unlikely or rare sequences – POS (different tagsets) – Tokens – Raw frequency vs. mutual information • Most publications are in the area of context-sensitive spelling correction – Real word errors – Resulting sentence can be grammatical 51 Test Corpus - Example • Missing Word Error She didn’t want to face him She didn’t to face him 52 Test Corpus – Example 2 • Context-sensitive spelling error I love them both I love then both 53 Cross-validation • • • • Standard deviation below 0.006 Except Method 4: 0.026 High number of test items Report average percentage 54 Example Run Stdev F-Score 1 0.654 2 0.655 3 0.655 4 0.655 5 0.653 6 0.652 7 0.653 8 0.657 9 0.654 10 0.653 Method 1 – Agreement errors: 65.4 % average F-Score 0.001 55 POS n-grams and Agreement Errors n = 2, 3, 4, 5 XLE parser F-Score 65 % Best Accuracy 55 % Best F-Score 66 % 56 POS n-grams and ContextSensitive Spelling Errors Best Accuracy 66 % XLE 60 % Best F-Score 69 % n = 2, 3, 4, 5 57 POS n-grams and Extra Word Errors Best Accuracy 68 % XLE 62 % Best F-Score 70 % n = 2, 3, 4, 5 58 POS n-grams and Missing Word Errors n = 2, 3, 4, 5 XLE 53 % Best Accuracy 59 % Best F-Score 67 % 59