Solving Arithmetic Word Problems Automatically Using Transformer and Unambiguous Representations

12/02/2019
by   Kaden Griffith, et al.
0

Constructing accurate and automatic solvers of math word problems has proven to be quite challenging. Prior attempts using machine learning have been trained on corpora specific to math word problems to produce arithmetic expressions in infix notation before answer computation. We find that custom-built neural networks have struggled to generalize well. This paper outlines the use of Transformer networks trained to translate math word problems to equivalent arithmetic expressions in infix, prefix, and postfix notations. In addition to training directly on domain-specific corpora, we use an approach that pre-trains on a general text corpus to provide foundational language abilities to explore if it improves performance. We compare results produced by a large number of neural configurations and find that most configurations outperform previously reported approaches on three of four datasets with significant increases in accuracy of over 20 percentage points. The best neural approaches boost accuracy by almost 10 compared to the previous state of the art.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2021

Solving Arithmetic Word Problems with Transformers and Preprocessing of Problem Text

This paper outlines the use of Transformer networks trained to translate...
research
08/04/2016

Solving General Arithmetic Word Problems

This paper presents a novel approach to automatically solving arithmetic...
research
07/24/2023

Explaining Math Word Problem Solvers

Automated math word problem solvers based on neural networks have succes...
research
11/22/2022

OLGA : An Ontology and LSTM-based approach for generating Arithmetic Word Problems (AWPs) of transfer type

Machine generation of Arithmetic Word Problems (AWPs) is challenging as ...
research
02/06/2014

An Autoencoder Approach to Learning Bilingual Word Representations

Cross-language learning allows us to use training data from one language...
research
09/15/2023

Using Large Language Model to Solve and Explain Physics Word Problems Approaching Human Level

Our work demonstrates that large language model (LLM) pre-trained on tex...
research
02/01/2018

Adapting predominant and novel sense discovery algorithms for identifying corpus-specific sense differences

Word senses are not static and may have temporal, spatial or corpus-spec...

Please sign up or login with your details

Forgot password? Click here to reset