A novel multi-layer feature fusion-based BERT-CNN for sentence representation learning and classification

Author:

Alyoubi Khaled Hamed,Alotaibi Fahd Saleh,Kumar Akhil,Gupta Vishal,Sharma Akashdeep

Abstract

Purpose The purpose of this paper is to describe a new approach to sentence representation learning leading to text classification using Bidirectional Encoder Representations from Transformers (BERT) embeddings. This work proposes a novel BERT-convolutional neural network (CNN)-based model for sentence representation learning and text classification. The proposed model can be used by industries that work in the area of classification of similarity scores between the texts and sentiments and opinion analysis. Design/methodology/approach The approach developed is based on the use of the BERT model to provide distinct features from its transformer encoder layers to the CNNs to achieve multi-layer feature fusion. To achieve multi-layer feature fusion, the distinct feature vectors of the last three layers of the BERT are passed to three separate CNN layers to generate a rich feature representation that can be used for extracting the keywords in the sentences. For sentence representation learning and text classification, the proposed model is trained and tested on the Stanford Sentiment Treebank-2 (SST-2) data set for sentiment analysis and the Quora Question Pair (QQP) data set for sentence classification. To obtain benchmark results, a selective training approach has been applied with the proposed model. Findings On the SST-2 data set, the proposed model achieved an accuracy of 92.90%, whereas, on the QQP data set, it achieved an accuracy of 91.51%. For other evaluation metrics such as precision, recall and F1 Score, the results obtained are overwhelming. The results with the proposed model are 1.17%–1.2% better as compared to the original BERT model on the SST-2 and QQP data sets. Originality/value The novelty of the proposed model lies in the multi-layer feature fusion between the last three layers of the BERT model with CNN layers and the selective training approach based on gated pruning to achieve benchmark results.

Publisher

Emerald

Reference45 articles.

1. Transfer fine-tuning of BERT with phrasal paraphrases;Computer Speech & Language,2020

2. An answer recommendation framework for an online cancer community forum;Multimedia Tools and Applications,2023

3. CNN-Trans-Enc: a cnn-enhanced transformer-encoder on top of static BERT representations for document classification,2022

4. Enriching word vectors with subword information,2016

5. Multilingual sentiment analysis: an RNN-based framework for limited data,2018

Cited by 8 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Robust Synthetic-to-Real Ensemble Dehazing Algorithm With the Intermediate Domain;IEEE Transactions on Computational Social Systems;2024-08

2. Unsupervised face image deblurring via disentangled representation learning;Pattern Recognition Letters;2024-07

3. Imitation Learning of Robotic Arm with Hierarchical Training Based on Human Videos;2024 39th Youth Academic Annual Conference of Chinese Association of Automation (YAC);2024-06-07

4. UTILIZATION OF ROBOTS IN INDUSTRIAL PRACTICE;MM Science Journal;2024-06-03

5. Self-Supervised 3-D Semantic Representation Learning for Vision-and-Language Navigation;IEEE Transactions on Neural Networks and Learning Systems;2024

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3