×
Home Current Archive Editorial board
Instructions for papers
For Authors Aim & Scope Contact
Original scientific article

A PMI-DRIVEN APPROACH WITH CONVENTIONAL BERT FOR OPTIMIZING TEXT SUMMARIZATION

By
R. Ramesh Orcid logo ,
R. Ramesh

Assistant Professor, Department of Computer Applications, Thanthai Periyar Government Arts & Sciences College , Trichy, Tamil Nadu , India

N. Subalakshmi Orcid logo ,
N. Subalakshmi

Assistant Professor, Department of Computer and Information Science, Annamalai University , Chidambaram, Tamil Nadu , India

S. Selvarani Orcid logo ,
S. Selvarani

Assistant Professor, Department of Computer Science, Alagappa Government Arts College , Karaikudi, Tamil Nadu , India

K. Kavitha Orcid logo ,
K. Kavitha

Assistant Professor (Selection Grade), Department of Electrical & Electronics Engineering, Annamalai University , Annamalai Nagar, Chidambaram, Tamil Nadu , India

M. Jeyakarthic Orcid logo
M. Jeyakarthic

Assistant Professor, Department of Computer and Information Science, Annamalai University , Annamalai Nagar, Chidambaram, Tamil Nadu , India

Abstract

Text summarization plays a crucial role in natural language processing by condensing large volumes of textual information into concise and meaningful summaries. With the rapid growth of digital content, existing summarization approaches often struggle to balance contextual understanding and semantic relevance. This paper presents a PMI-driven BERT-based text summarization framework that integrates Pointwise Mutual Information (PMI) as a statistical pre-processing mechanism with a fine-tuned Conventional BERT model to enhance summary quality. PMI is employed to identify and rank semantically significant terms based on co-occurrence patterns, enabling effective keyword and phrase prioritization before summarization. The ranked textual representation is then processed using a summarization-specific decoder layer added on top of the BERT encoder to generate coherent and context-aware summaries. The proposed framework is evaluated on the CNN/Daily Mail dataset comprising over 300,000 news articles, using ROUGE-1, ROUGE-2, and ROUGE-L metrics for performance assessment. Experimental results demonstrate that the proposed method achieves ROUGE-1, ROUGE-2, and ROUGE-L scores of 46.9, 27.61, and 45.68 respectively, outperforming baseline models such as Seq2Seq, Seq2Sick, and Prefix-Tuning by an average margin of 2–3%. The experiments were conducted using Python with the PyTorch deep learning framework on a CPU-based environment. The results indicate that PMI-based pre-processing significantly improves contextual relevance and semantic consistency in generated summaries. This framework demonstrates robustness and scalability, making it suitable for large-scale document summarization tasks.

References

1.
Belwal RC, Gupta A. Automatic text summarization techniques: categorization and contemporary challenges. Information Processing and Management. 2025;62(2):103612.
2.
Aswani S, Choudhary K, Shetty S, Nur N. Automatic text summarization of scientific articles using transformers—A brief review. Journal of Autonomous Intelligence. 2024;7(5).
3.
Wibawa AP, Kurniawan F. A survey of text summarization: Techniques, evaluation and challenges. Natural Language Processing Journal. 2024 Jun 1;7:100070. .
4.
Zhang Y, Jin H, Meng D, Wang J, Tan J. A comprehensive survey on automatic text summarization with exploration of LLM-based methods. Neurocomputing. 2025 Nov 3:131928.
5.
Liu W, Sun Y, Yu B, Wang H, Peng Q, Hou M, Guo H, Wang H, Liu C. Automatic text summarization method based on improved Text Rank algorithm and K-means clustering. Knowledge-Based Systems. 2024 Mar 5;287:111447.

Citation

This is an open access article distributed under the  Creative Commons Attribution Non-Commercial License (CC BY-NC) License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. 

Article metrics

Google scholar: See link

The statements, opinions and data contained in the journal are solely those of the individual authors and contributors and not of the publisher and the editor(s). We stay neutral with regard to jurisdictional claims in published maps and institutional affiliations.