Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers

We assessed whether constructing a mathematical knowledge graph for a knowledge question-answering system or a course recommendation system, Named Entity Recognition (NER), is indispensable. The accuracy of its recognition directly affects the actual performance of these subsequent tasks. In order t...

Full description

Saved in:
Bibliographic Details
Main Authors: Wei Song, He Zheng, Shuaiqi Ma, Mingze Zhang, Wei Guo, Keqing Ning
Format: Article
Language:English
Published: MDPI AG 2025-01-01
Series:Information
Subjects:
Online Access:https://www.mdpi.com/2078-2489/16/1/42
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832588333051543552
author Wei Song
He Zheng
Shuaiqi Ma
Mingze Zhang
Wei Guo
Keqing Ning
author_facet Wei Song
He Zheng
Shuaiqi Ma
Mingze Zhang
Wei Guo
Keqing Ning
author_sort Wei Song
collection DOAJ
description We assessed whether constructing a mathematical knowledge graph for a knowledge question-answering system or a course recommendation system, Named Entity Recognition (NER), is indispensable. The accuracy of its recognition directly affects the actual performance of these subsequent tasks. In order to improve the accuracy of mathematical knowledge entity recognition and provide effective support for subsequent functionalities, this paper adopts the latest pre-trained language model, LERT, combined with a Bidirectional Gated Recurrent Unit (BiGRU), Iterated Dilated Convolutional Neural Networks (IDCNNs), and Conditional Random Fields (CRFs), to construct the LERT-BiGRU-IDCNN-CRF model. First, LERT provides context-related word vectors, and then the BiGRU captures both long-distance and short-distance information, the IDCNN retrieves local information, and finally the CRF is decoded to output the corresponding labels. Experimental results show that the accuracy of this model when recognizing mathematical concepts and theorem entities is 97.22%, the recall score is 97.47%, and the F1 score is 97.34%. This model can accurately recognize the required entities, and, through comparison, this method outperforms the current state-of-the-art entity recognition models.
format Article
id doaj-art-30e59fd093054459864e8e180400580a
institution Kabale University
issn 2078-2489
language English
publishDate 2025-01-01
publisher MDPI AG
record_format Article
series Information
spelling doaj-art-30e59fd093054459864e8e180400580a2025-01-24T13:35:15ZengMDPI AGInformation2078-24892025-01-011614210.3390/info16010042Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from TransformersWei Song0He Zheng1Shuaiqi Ma2Mingze Zhang3Wei Guo4Keqing Ning5School of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaState Grid Jilin Electric Power Research Institute, Changchun 130015, ChinaSchool of Electrical and Control Engineering, North China University of Technology, Beijing 100144, ChinaSchool of Information Science and Technology, North China University of Technology, Beijing 100144, ChinaWe assessed whether constructing a mathematical knowledge graph for a knowledge question-answering system or a course recommendation system, Named Entity Recognition (NER), is indispensable. The accuracy of its recognition directly affects the actual performance of these subsequent tasks. In order to improve the accuracy of mathematical knowledge entity recognition and provide effective support for subsequent functionalities, this paper adopts the latest pre-trained language model, LERT, combined with a Bidirectional Gated Recurrent Unit (BiGRU), Iterated Dilated Convolutional Neural Networks (IDCNNs), and Conditional Random Fields (CRFs), to construct the LERT-BiGRU-IDCNN-CRF model. First, LERT provides context-related word vectors, and then the BiGRU captures both long-distance and short-distance information, the IDCNN retrieves local information, and finally the CRF is decoded to output the corresponding labels. Experimental results show that the accuracy of this model when recognizing mathematical concepts and theorem entities is 97.22%, the recall score is 97.47%, and the F1 score is 97.34%. This model can accurately recognize the required entities, and, through comparison, this method outperforms the current state-of-the-art entity recognition models.https://www.mdpi.com/2078-2489/16/1/42mathematical knowledge entity recognitionLERTBiGRUIDCNNCRF
spellingShingle Wei Song
He Zheng
Shuaiqi Ma
Mingze Zhang
Wei Guo
Keqing Ning
Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers
Information
mathematical knowledge entity recognition
LERT
BiGRU
IDCNN
CRF
title Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers
title_full Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers
title_fullStr Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers
title_full_unstemmed Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers
title_short Chinese Mathematical Knowledge Entity Recognition Based on Linguistically Motivated Bidirectional Encoder Representation from Transformers
title_sort chinese mathematical knowledge entity recognition based on linguistically motivated bidirectional encoder representation from transformers
topic mathematical knowledge entity recognition
LERT
BiGRU
IDCNN
CRF
url https://www.mdpi.com/2078-2489/16/1/42
work_keys_str_mv AT weisong chinesemathematicalknowledgeentityrecognitionbasedonlinguisticallymotivatedbidirectionalencoderrepresentationfromtransformers
AT hezheng chinesemathematicalknowledgeentityrecognitionbasedonlinguisticallymotivatedbidirectionalencoderrepresentationfromtransformers
AT shuaiqima chinesemathematicalknowledgeentityrecognitionbasedonlinguisticallymotivatedbidirectionalencoderrepresentationfromtransformers
AT mingzezhang chinesemathematicalknowledgeentityrecognitionbasedonlinguisticallymotivatedbidirectionalencoderrepresentationfromtransformers
AT weiguo chinesemathematicalknowledgeentityrecognitionbasedonlinguisticallymotivatedbidirectionalencoderrepresentationfromtransformers
AT keqingning chinesemathematicalknowledgeentityrecognitionbasedonlinguisticallymotivatedbidirectionalencoderrepresentationfromtransformers