원문정보
A Case Study of Inter-rater Comparability in Interpreting Assessment.
초록
영어
This paper attempts to explore issues of comparability in the assessment of interpreting when more than two raters are involved. It is often argued that the assessment results are not reliable as raters apply different criteria and rating systems. It is complicated, however, to investigate the issues as they are mostly stem from the raters' subconscious; indeed, the raters themselves may find it difficult to specify these issues. In order to investigate the raters' internal systems, a series of assessments and coordination sessions have been devised, ultimately aiming to achieve comparability among raters. While the raters all inevitably develop different assessment criteria due to their different work experiences and educational backgrounds, they could be guided to share a common set of criteria for assessment through a series of assessment exercises. This may be particularly helpful for novice teachers, not only for their assessment but also for their teaching. Further, these coordination activities could contribute to coordination of the implementation of the curriculum when multiple classes are run in parallel.
목차
1. 서론
2. 이론적 배경
2.1. 통역 평가 기준
2.2. 평가자 요인
3. 사례 연구
3.1. 사례 배경
3.2. 자료 수집
4. 분석 결과
4.1. 평가자 별 기준 비교
4.2. 평가자 간 비교가능성 분석
5. 결론
참고문헌
부록