研究生: |
朱安安 Chu, An-An |
---|---|
論文名稱: |
中翻英自動評分系統實作 An Automated Scoring System for Chinese-to-English Translation |
指導教授: |
張智星
Jang, Jyh-Shing Roger 張俊盛 Chang, Jason S. |
口試委員: |
張智星
張俊盛 陳浩然 |
學位類別: |
碩士 Master |
系所名稱: |
電機資訊學院 - 資訊系統與應用研究所 Institute of Information Systems and Applications |
論文出版年: | 2012 |
畢業學年度: | 100 |
語文別: | 中文 |
論文頁數: | 52 |
中文關鍵詞: | 自動評分 、機器翻譯 、電腦輔助學習 |
外文關鍵詞: | METEOR, CALL, MT evaluation |
相關次數: | 點閱:3 下載:0 |
分享至: |
查詢本校圖書館目錄 查詢臺灣博碩士論文知識加值系統 勘誤回報 |
本研究實作「中翻英自動評分系統」,並驗證自動評分可應用於大規模考試(如:大學學測、指考)的中翻英翻譯題批改,作為人工閱卷的輔助。
在研究的第一階段,我們利用「METEOR機器翻譯評測指標」的評分方法實作中翻英翻譯題自動評分系統。
在研究的第二階段,我使用前階段製作的評分系統批改學生翻譯,隨後計算系統評分與教師評分的相關度,結果顯示:自動評分的結果與真人教師的判斷具有高度相關性,因此足以做為閱卷的輔助。
在研究的第三階段,我們利用前階段取得的數據檢視研究結果,並討論可行的改良方法,希望經過進一步的測試與優化後,我們的系統可實際應用於正式考試,提昇閱卷效率與評分品質。
We built an automated scoring system for Chinese-to-English translation, and proved that this system is capable of assisting English teachers in grading translations.
First of all, we designed our automated scoring system based on the scoring methods of METEOR (an MT evaluation metric).
Secondly, we collected student translations and graded these translations with our system; meanwhile, we also invited English experts to grade these translations. After comparing the system’s and experts’ scorings, we concluded that the system’s performance is highly correlated to the experts’.
In the last chapter of this thesis, we discussed how to optimize our system in the future. We hope our research achievements can benefit some large-scale examinations such as university entrance exams, making grading papers more efficiently and objectively.
[1] Michael Denkowski and Alon Lavie, "Meteor 1.3: Automatic Metric for Reliable Optimization and Evaluation of Machine Translation Systems", Proceedings of the EMNLP 2011 Workshop on Statistical Machine Translation, 2011.
[2] Alon Lavie and Abhaya Agarwal, "METEOR: An Automatic Metric for MT Evaluation with High Levels of Correlation with Human Judgments", Proceedings of the ACL 2007 Workshop on Statistical Machine Translation, 2007.
[3] M. Snover, B. Dorr, R. Schwartz, L. Micciulla, and J. Makhoul. "A study of translation edit rate with targeted human annotation", In 7th Association for Machine Translation in the America, pages 223–231, Cambridge, MA. 2006.
[4] Dikli, S. "An Overview of Automated Scoring of Essays. Journal of Technology", Learning, and Assessment, 5(1). Retrieved [date] from http://www.jtla.org. 2006.
[5] Satanjeev Banerjee and Alon Lavie, "METEOR: An Automatic Metric for MT Evaluation with Improved Correlation with Human Judgments", Proceedings of the ACL 2005 Workshop on Intrinsic and Extrinsic Evaluation Measures for MT and/or Summarization, 2005.
[6] Chin-Yew Lin, Eduard Hovy, "Automatic evaluation of summaries using N-gram co-occurrence statistics", Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology, p.71-78, May 27-June 01, 2003.
[7] Kishore Papineni, Salim Roukos, Todd Ward and WeiJing Zhu. "BLEU: a Method for Automatic Evaluation of Machine Translation. " In Proceedings of the 40th Annual Meeting of the Association for Computational Linguistics (ACL-02). Philadelphia, PA. July 2002. pp. 311-318.
[8] Python, http://python.org/
[9] 王茜,“基於BLEU的英語翻譯自動評分研究”,安徽電子資訊職業技術學院學報2010年04期
[10] 張俊盛,“電腦輔助閱卷的條件與時機”,科學人2009年2月號第84期
[11] 王 雷、常寶寶,“大學英語翻譯考試人工輔助電腦評分初探”,外語電化教學,2009
[12] 嚴可、胡國平、魏思、戴禮榮,“面向大規模英語口語機考的複述題自動評分技術”,清華大學學報(自然科學版),2009年第S1期
[13] 嚴可,“英文背誦題及複述題自動評測技術研究”,中國科學技術大學碩士論文,2009年
[14] 李俊毅,“語音評分”,國立清華大學碩士論文,2002年