Machine translation translates a text from one language to another, while text simplification converts a text from its original form to a simpler one, usually in the same language. This survey paper discusses the evaluation (manual and automatic) of both fields, providing an overview of existing metrics along with their strengths and weaknesses. The first chapter takes an in-depth look at machine translation evaluation metrics, namely BLEU, NIST, AMBER, LEPOR, MP4IBM1, TER, MMS, METEOR, TESLA, RTE, and HTER. The second chapter focuses more generally on text simplification, starting with a discussion of the theoretical underpinnings of the field (i.e what ``simple'' means). Then, an overview of automatic evaluation metrics, namely BLEU and Flesch-Kincaid, is given, along with common approaches to text simplification. The paper concludes with a discussion of the future trajectory of both fields.
Identifer | oai:union.ndltd.org:CLAREMONT/oai:scholarship.claremont.edu:scripps_theses-1766 |
Date | 01 January 2016 |
Creators | Tapkanova, Elmira |
Publisher | Scholarship @ Claremont |
Source Sets | Claremont Colleges |
Detected Language | English |
Type | text |
Format | application/pdf |
Source | Scripps Senior Theses |
Rights | © 2015 Elmira Tapkanova, default |
Page generated in 0.0017 seconds