A Generalization of Compositional Matrix-Space Models of Language for Short Answer Scoring

Authors

  • STEPHEN HUAN
  • DHRUV SUNDARARAMAN
  • SHAWN MALIK
  • Mihai Boicu

DOI:

https://doi.org/10.13021/jssr2020.3146

Abstract

Automated Short Answer Grading (ASAG) consists of using computer programs to grade student short responses to free response questions. Current state of the art algorithms for ASAG rely on statistical representations of words, commonly generated through deep learning techniques like word2vec. One property of word2vec is commutativity, i.e. switching the order of words does not change the embedding. Motivated by this shortcoming, word2mat embeds words as matrices, whose multiplication is noncommutative. However, word2mat is both linear and associative, while English has nonlinear effects in certain contexts. To address these mathematical properties, we propose representing words as shallow neural networks. We then generalize word2mat by introducing a simple nonlinear activation function between each matrix multiplication, which we show makes the embedding noncommutative, nonlinear, and non-associative. Empirically, the nonlinear model performs an average of 0.83% worse than word2mat on the SentEval framework, and a neural network trained on both models’ embeddings to predict student grades has nearly identical accuracy. Possible justifications of the discrepancy between theoretical advantages and empirical performance are computational power and limited training time. 

Published

2022-12-13

Issue

Section

College of Engineering and Computing: Department of Information Sciences and Technology

Categories