TY - JOUR
T1 - A hierarchical classification approach to automated essay scoring
AU - McNamara, Danielle S.
AU - Crossley, Scott A.
AU - Roscoe, Rod D.
AU - Allen, Laura K.
AU - Dai, Jianmin
N1 - Publisher Copyright:
© 2014 Elsevier Ltd.
PY - 2015/1/1
Y1 - 2015/1/1
N2 - This study evaluates the use of a hierarchical classification approach to automated assessment of essays. Automated essay scoring (AES) generally relies on machine learning techniques that compute essay scores using a set of text variables. Unlike previous studies that rely on regression models, this study computes essay scores using a hierarchical approach, analogous to an incremental algorithm for hierarchical classification. The corpus in this study consists of 1243 argumentative (persuasive) essays written on 14 different prompts, across 3 different grade levels (9th grade, 11th grade, college freshman), and four different time limits for writing or temporal conditions (untimed essays and essays written in 10, 15, and 25. minute increments). The features included in the analysis are computed using the automated tools, Coh-Metrix, the Writing Assessment Tool (WAT), and Linguistic Inquiry and Word Count (LIWC). Overall, the models developed to score all the essays in the data set report 55% exact accuracy and 92% adjacent accuracy between the predicted essay scores and the human scores. The results indicate that this is a promising approach to AES that could provide more specific feedback to writers and may be relevant to other natural language computations, such as the scoring of short answers in comprehension or knowledge assessments.
AB - This study evaluates the use of a hierarchical classification approach to automated assessment of essays. Automated essay scoring (AES) generally relies on machine learning techniques that compute essay scores using a set of text variables. Unlike previous studies that rely on regression models, this study computes essay scores using a hierarchical approach, analogous to an incremental algorithm for hierarchical classification. The corpus in this study consists of 1243 argumentative (persuasive) essays written on 14 different prompts, across 3 different grade levels (9th grade, 11th grade, college freshman), and four different time limits for writing or temporal conditions (untimed essays and essays written in 10, 15, and 25. minute increments). The features included in the analysis are computed using the automated tools, Coh-Metrix, the Writing Assessment Tool (WAT), and Linguistic Inquiry and Word Count (LIWC). Overall, the models developed to score all the essays in the data set report 55% exact accuracy and 92% adjacent accuracy between the predicted essay scores and the human scores. The results indicate that this is a promising approach to AES that could provide more specific feedback to writers and may be relevant to other natural language computations, such as the scoring of short answers in comprehension or knowledge assessments.
KW - AES
KW - Automated essay scoring
KW - Hierarchical classification
KW - Writing assessment
UR - http://www.scopus.com/inward/record.url?scp=84911914341&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84911914341&partnerID=8YFLogxK
U2 - 10.1016/j.asw.2014.09.002
DO - 10.1016/j.asw.2014.09.002
M3 - Article
AN - SCOPUS:84911914341
SN - 1075-2935
VL - 23
SP - 35
EP - 59
JO - Assessing Writing
JF - Assessing Writing
ER -