Given a question, a reference answer, and the answer given by the student, the aim of the automatic short answer grading task is to assign a grade to the student’s answer. We use for this a large number of matching rules relying on recognizing entailment relation between dependency structures of the two answers. Comparison of the grades generated by our method with those given by human judges on a computer science dataset shows a quite promising maximum correlation of 0.627.
MohlerM. and MihalceaR., Text-to-text semantic similarity for automatic short answer grading, EACL: 12th Conference of the European Chapter of the ACL, (2009), pp. 567–575.
2.
MohlerM., BunescuR. and MihalceaR., Learning to grade short answer questions using semantic similarity measures and dependency graph alignments, HLT: 49th Annual Meeting of the ACL: Human Language Technologies, vol. 1, 2011, pp. 752–762.
3.
GommaW.H. and FahmyA.A., Short answer grading using string similarity and corpus-based similarity, International Journal of Advanced Computer Science and Applications3(11) (2012).
4.
AlotaibiS.T. and MirzaA.A., Hybrid approach for automatic short answer marking, SWDSI: 43rd Annual Meeting of the Southwest Decision Sciences Institute, 2012.
5.
PulmanS.G. and SukkariehJ.Z., Automatic short answer marking, EdAppsNLP:2ndWorkshop on Building Educational Applications Using NLP, 2005, pp. 9–16.
6.
SultanM.A., SalazarC. and SumnerT., Fast and easy short answer grading high accuracy, NAACL-HLT, 2016, pp. 1070–1075.
7.
PerezD., AlfonsecaE. and RodriguezP., Application of the BLEU method for evaluating free-text answers in an e-learning environment, LREC: Language Resources and Evaluation Conference, 2004.
8.
SelviP. and BanerjeeA.K., Automatic short-answer grading system (ASAGS), InterJRI Computer Science and Networking2(1) (2010).
9.
BachmanL.F., CarrN., KameiG., KimM., PanM.J., SalvadorC. and SawakiY., A reliable approach to automatic assessment of short answer free responses, COLING: 19th International Conference on Computational Linguistics2 (2002), 1–4.
10.
SukkariehJ.Z. and BlackmoreJ., C-rater: Automatic content scoring for short constructed responses, 22nd International FLAIRS Conference, 2009, pp. 290–295.
11.
BaileyS. and MeurersD., Diagnosing meaning errors in short answers to reading comprehension questions, 3rd ACL Workshop on Innovative Use of NLP for Building Educational Applications, 2008, pp. 107–115.
12.
BasuS., JacobsC. and VanderwendeL., Powergrading: A clustering approach to amplify human effort for short answer grading, Transactions of the ACL, 2013, pp. 391–402.
13.
RoyS., DandapatS., NageshA. and NarahariY., Wisdom of students: A consistent automatic short answer grading technique, 13th Intl Conference on Natural Language Processing2016, pp. 178–187.
14.
BasakR., NaskarS.K. and GelbukhA., ALexico-syntactic-semantic approach for recognizing textual entailment, https://goo.gl/KsbHRs.