May 9, 2016 · This paper describes improvements made to the GLEU metric that address problems that arise when using an increasing number of reference sets.
This work proposes MAEGE, an automatic methodology for GEC metric validation, that overcomes many of the difficulties in the existing methodology.
May 9, 2016 · GLEU (Generalized Language Understanding. Evaluation)2 was designed and developed using two sets of annotations as references, with a tun- able ...
May 9, 2016 · This paper describes improvements made to the GLEU metric that address problems that arise when using an increasing number of reference sets.
May 10, 2016 · The changes to GLEU and updated results to our ACL 2015 paper are described in the eprint, GLEU Without Tuning. The citation for the updated ...
Additionally, GLEU penalizes n-grams in the hypotheses that match part of the input but not the reference. The original formulation (Napoles et al.,. 2015) ...
The GLEU metric was proposed for evaluating grammatical error corrections using n-gram overlap with a set of reference sentences, as opposed to precision/recall ...
GLEU: "computes n-gram precisions over the reference but assigns more weight to n-grams that have been correctly changed from the source." Python code. WER ( ...
Jul 15, 2024 · This paper proposes SEEDA, a new dataset for GEC meta-evaluation. SEEDA consists of corrections with human ratings along two different granularities.
Gleu without tuning. arXiv. 784 preprint arXiv:1605.02592. 785. Courtney Napoles, Keisuke Sakaguchi, and Joel. 786. Tetreault. 2017. Jfleg: A fluency corpus and ...