Pillars of Grammatical Error Correction: Comprehensive Inspection Of Contemporary Approaches In The Era of Large Language Models
Paper
•
2404.14914
•
Published
ONNX quantized version of Grammarly's GECToR 2024 model for browser-based grammatical error correction with Transformers.js.
GECToR uses a token classification approach - instead of generating corrected text, it predicts edit operations for each token:
$KEEP - Keep token unchanged$DELETE - Remove token$REPLACE_word - Replace with specific word$APPEND_word - Append word after token$TRANSFORM_* - Apply transformation (case, verb form, etc.)The model runs iteratively (typically 2-3 passes) until no more edits are predicted.
import { pipeline } from '@huggingface/transformers';
const classifier = await pipeline(
'token-classification',
'YOUR_USERNAME/gector-large-2024',
{ dtype: 'q8' }
);
const result = await classifier('He go to school yesterday.');
// Returns token predictions with edit tags
Best accuracy among GECToR variants. Recommended for quality-critical applications.
Apache 2.0 (following original model license)