TY - JOUR
T1 - Summarization of scholarly articles using BERT and BiGRU
T2 - Deep learning-based extractive approach
AU - Bano, Sheher
AU - Khalid, Shah
AU - Tairan, Nasser Mansoor
AU - Shah, Habib
AU - Khattak, Hasan Ali
PY - 2023/10
Y1 - 2023/10
N2 - Extractive text summarization involves selecting and combining key sentences directly from the original text, rather than generating new content. While various methods, both statistical and graph-based, have been explored for this purpose, accurately capturing the intended meaning remains a challenge. To address this, researchers are investigating innovative techniques that harness deep learning models like BERT (Bidirectional Encoder Representations from Transformers). However, BERT has limitations in summarizing lengthy documents due to input length constraints. To find a more effective solution, we propose a novel approach. This approach combines the power of BERT, a transformer network pre-trained on extensive self-supervised datasets, with BiGRU (Bidirectional Gated Recurrent Units), a recurrent neural network that captures sequential dependencies within the text for extracting salient information. Our method involves using BERT to generate sentence-level embeddings, which are then fed into the BiGRU network. This allows us to achieve a comprehensive understanding of the complete document's context. In experimental analysis conducted on arXiv and PubMed datasets, the proposed approach outperformed several state-of-the-art models. It achieved remarkable ROUGE-F scores of (46.7, 19.4, 35.4) and (47.0, 21.3, 39.7) on these datasets respectively. The proposed fusion of BERT and BiGRU significantly enhances extractive text summarization. It shows promising potential for summarizing lengthy documents and benefiting various domains that require concise and informative summaries.
AB - Extractive text summarization involves selecting and combining key sentences directly from the original text, rather than generating new content. While various methods, both statistical and graph-based, have been explored for this purpose, accurately capturing the intended meaning remains a challenge. To address this, researchers are investigating innovative techniques that harness deep learning models like BERT (Bidirectional Encoder Representations from Transformers). However, BERT has limitations in summarizing lengthy documents due to input length constraints. To find a more effective solution, we propose a novel approach. This approach combines the power of BERT, a transformer network pre-trained on extensive self-supervised datasets, with BiGRU (Bidirectional Gated Recurrent Units), a recurrent neural network that captures sequential dependencies within the text for extracting salient information. Our method involves using BERT to generate sentence-level embeddings, which are then fed into the BiGRU network. This allows us to achieve a comprehensive understanding of the complete document's context. In experimental analysis conducted on arXiv and PubMed datasets, the proposed approach outperformed several state-of-the-art models. It achieved remarkable ROUGE-F scores of (46.7, 19.4, 35.4) and (47.0, 21.3, 39.7) on these datasets respectively. The proposed fusion of BERT and BiGRU significantly enhances extractive text summarization. It shows promising potential for summarizing lengthy documents and benefiting various domains that require concise and informative summaries.
KW - Attention mechanism
KW - BERT
KW - BiGRU
KW - Text summarization
UR - https://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=web_of_science_starterapi&SrcAuth=WosAPI&KeyUT=WOS:001076888300001&DestLinkType=FullRecord&DestApp=WOS_CPL
UR - http://www.scopus.com/inward/record.url?scp=85171344529&partnerID=8YFLogxK
U2 - 10.1016/j.jksuci.2023.101739
DO - 10.1016/j.jksuci.2023.101739
M3 - Article
SN - 1319-1578
VL - 35
JO - Journal of King Saud University Computer and Information Sciences
JF - Journal of King Saud University Computer and Information Sciences
IS - 9
M1 - 101739
ER -