Short text automatic scoring system based on BERT-BiLSTM model

Aiming at the problems of sparse features, polysemy of one word and less context related information in short text automatic scoring, a short text automatic scoring model based on bidirectional encoder representations from transformers - bidirectional long short-term memory (BERT-BiLSTM) is proposed...

Full description

Saved in:
Bibliographic Details
Main Authors: XIA Linzhong, YE Jianfeng, LUO De’an, GUAN Mingxiang, LIU Jun, CAO Xuemei
Format: Article
Language:English
Published: Science Press (China Science Publishing & Media Ltd.) 2022-05-01
Series:Shenzhen Daxue xuebao. Ligong ban
Subjects:
Online Access:https://journal.szu.edu.cn/en/#/digest?ArticleID=2434
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Aiming at the problems of sparse features, polysemy of one word and less context related information in short text automatic scoring, a short text automatic scoring model based on bidirectional encoder representations from transformers - bidirectional long short-term memory (BERT-BiLSTM) is proposed. Firstly, the large-scale corpus is pre-trained with bidirectional encoder representations from transformers (BERT) language model to acquire the semantic features of the general language. Then the semantic features of short text and the semantics of keywords in a specific context are acquired through the short text data for the pre-fine tuning downstream specific tasks set pre-fined by BERT. And then the deep-seated context dependency is captured through bidirectional long short-term memory (BiLSTM). Finally, the obtained feature vectors are input into Softmax regression model for automatic scoring. The experimental results show that compared with other benchmark models of convolutional neural networks(CNN), character-level CNN (CharCNN), long short-term memory (LSTM) and BERT, the short text automatic scoring model based on BERT-BiLSTM achieves the best average value of quadratic weighted kappa coefficient.
ISSN:1000-2618