Publication Type
Conference Paper
Version
acceptedVersion
Publication Date
8-2023
Abstract
Existing generative pre-trained language models (e.g., GPT) focus on modeling the language structure and semantics of general texts. However, those models do not consider the numerical properties of numbers and cannot perform robustly on numerical reasoning tasks (e.g., math word problems and measurement estimation). In this paper, we propose NumGPT, a generative pre-trained model that explicitly models the numerical properties of numbers in texts. Specifically, it leverages a prototype-based numeral embedding to encode the mantissa of the number and an individual embedding to encode the exponent of the number. A numeral-aware loss function is designed to integrate numerals into the pre-training objective of NumGPT. We conduct extensive experiments on four different datasets to evaluate the numeracy ability of NumGPT. The experiment results show that NumGPT outperforms baseline models (e.g., GPT and GPT with DICE) on a range of numerical reasoning tasks such as measurement estimation, number comparison, math word problems, and magnitude classification. Ablation studies are also conducted to evaluate the impact of pre-training and model hyperparameters on the performance.
Discipline
Artificial Intelligence and Robotics
Research Areas
Intelligent Systems and Optimization
Publication
International Symposium on Large Language Models for Financial Services @ IJCAI 2023, Macau, August 20
Identifier
10.48550/arXiv.2109.03137
Publisher
Elsevier
City or Country
Macao
Citation
JIN, Zhihua; JIANG, Xin; WANG, Xiangbo; LIU, Qun; WANG, Yong; REN, Xiaozhe; and QU, Huamin.
NumGPT: Improving numeracy ability of generative pre-trained models. (2023). International Symposium on Large Language Models for Financial Services @ IJCAI 2023, Macau, August 20.
Available at: https://ink.library.smu.edu.sg/sis_research/8599
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://arxiv.org/abs/2109.03137