Timezone: »
Since the introduction of the original BERT (i.e., BASE BERT), researchers have developed various customized BERT models with improved performance for specific domains and tasks by exploiting the benefits of transfer learning. Due to the nature of mathematical texts, which often use domain specific vocabulary along with equations and math symbols, we posit that the development of a new BERT model for mathematics would be useful for many mathematical downstream tasks. In this paper, we introduce our multi-institutional effort (i.e., two learning platforms and three academic institutions in the US) toward this need: MathBERT, a model created by pre-training the BASE BERT model on a large mathematical corpus ranging from pre-kindergarten (pre-k), to high-school, to college graduate level mathematical content. In addition, we select three general NLP tasks that are often used in mathematics education: prediction of knowledge component, auto-grading open-ended Q&A, and knowledge tracing, to demonstrate the superiority of m over BASE BERT. Our experiments show that MathBERT outperforms prior best methods by 1.2-22% and BASE BERT by 2-8% on these tasks. In addition, we build a mathematics specific vocabulary mathVocab to train with MathBERT. We release MathBERT for public usage at: https://github.com/tbs17/MathBERT.
Author Information
Tracy Jia Shen (The Pennsylvania State University)
Michiharu Yamashita (The Pennsylvania State University)
Ethan Prihar (Worcester Polytechnic Institute)
Neil Heffernan (ASSISTments.org)
Xintao Wu (University of Arkansas)
Dr. Xintao Wu is the professor and the Charles D. Morgan/Acxiom Endowed Graduate Research Chair in Database and leads Social Awareness and Intelligent Learning (SAIL) Lab in Computer Science and Computer Engineering Department at University of Arkansas.
Ben Graff (Stride, Inc)
Dongwon Lee (Penn State University)
More from the Same Authors
-
2021 : Poster: Achieving Counterfactual Fairness for Causal Bandit »
Xintao Wu · Wen Huang · Lu Zhang -
2021 : MathBERT: A Pre-trained Language Model for General NLP Tasks in Mathematics Education »
Tracy Jia Shen -
2021 : Poster Session 1 »
Jiaqi Chen · Tanglin Xia · Sean Welleck · Jiacheng Liu · Ran Gong · Shifeng Huang · Wei Yu · Tracy Jia Shen -
2021 : Achieving Counterfactual Fairness for Causal Bandit »
Wen Huang · Lu Zhang · Xintao Wu -
2020 Poster: Fair Multiple Decision Making Through Soft Interventions »
Yaowei Hu · Yongkai Wu · Lu Zhang · Xintao Wu -
2019 : Poster Session »
Ayse Cakmak · Yunkai Zhang · Srijith Prabhakarannair Kusumam · Mohamed Osama Ahmed · Xintao Wu · Jayesh Choudhari · David I Inouye · Thomas Taylor · Michel Besserve · Ali Caner Turkmen · Kazi Islam · Antonio Artés · Amrith Setlur · Zhanghua Fu · Zhen Han · Abir De · Nan Du · Pablo Sanchez-Martin -
2019 : Insider Threat Detection via Hierarchical Neural Temporal Point Processes »
Xintao Wu -
2019 Poster: PC-Fairness: A Unified Framework for Measuring Causality-based Fairness »
Yongkai Wu · Lu Zhang · Xintao Wu · Hanghang Tong