earticle

논문검색

번역의 자동평가 : 기계번역 평가를 인간번역 평가에 적용해보기

원문정보

Automatic Evaluation of Human Translation.

정혜연

피인용수 : 0(자료제공 : 네이버학술정보)

초록

영어

‘Appropriate’ evaluation of translation quality requires three criteria: validity, reliability and practicality. The human translation evaluation can be considered as ‘valid’, if it is conducted by a professional evaluator, but it leaves out the other two criteria of evaluating human translation. Automatic evaluation, on the other hand, fulfills the second and the third, but misses the first criteria. It would be therefore, desirable, if both methods could complement each other. This paper attempts to verify this possibility. Using an online tool (interactive BLEU score evaluator), 150 pairs of translations were automatically evaluated and compared to the ratings of a human evaluator (professional translator and evaluator). The results showed only a slight inter-rater agreement (Kappa = .054, .109), but made us some implicit suggestions about the improvement of Kappa score. This can be increased, if calculations are based on 1-gram matches(or even morphemes), instead of 1-4 gram matches, and if the reference translations contain a certain quantity of synonyms. With regard to the validity of BLEU, Brevity Penalty should be substituted by a Length Penalty (Han et al., 2012) accordingly to the characteristics of human translation.

목차


1. 들어가며
2. 인간평가
3. 자동평가
3.1. 자동평가의 종류
3.2. BLEU
4. 실험
4.1. 번역사와 텍스트
4.2. 분석방식
5. 결과 및 토론
5.1. 일치도 세부 결과
5.2. BLEU 평가의 타당성
6. 나가며
참고문헌

저자정보

  • 정혜연 Chung, Hye-Yeon.. 한국외국어대학교

참고문헌

자료제공 : 네이버학술정보

    함께 이용한 논문

      ※ 기관로그인 시 무료 이용이 가능합니다.

      • 6,000원

      0개의 논문이 장바구니에 담겼습니다.