Semantic evaluation of machine translation.

Tak Ming WONG

    Research output: Conference PapersRGC 32 - Refereed conference paper (without host publication)peer-review

    Abstract

    It is recognized that many evaluation metrics of machine translation in use that focus on surface word level suffer from their lack of tolerance of linguistic variance, and the incorporation of linguistic features can improve their performance. To this end, WordNet is therefore widely utilized by recent evaluation metrics as a thesaurus for identifying synonym pairs. On this basis, word pairs in similar meaning, however, are still neglected. We investigate the significance of this particular word group to the performance of evaluation metrics. In our experiments we integrate eight different measures of lexical semantic similarity into an evaluation metric based on standard measures of unigram precision, recall and F-measure. It is found that a knowledge-based measure proposed by Wu and Palmer and a corpus-based measure, namely Latent Semantic Analysis, lead to an observable gain in correlation with human judgments of translation quality, in an extent to which comparable to the use of WordNet for synonyms.
    Original languageEnglish
    Publication statusPublished - 17 May 2010
    Event7th Language Resources and Evaluation Conference (LREC 2010) - Valletta, Malta
    Duration: 17 May 201023 May 2010

    Conference

    Conference7th Language Resources and Evaluation Conference (LREC 2010)
    PlaceMalta
    CityValletta
    Period17/05/1023/05/10

    Fingerprint

    Dive into the research topics of 'Semantic evaluation of machine translation.'. Together they form a unique fingerprint.

    Cite this