DSpace Repository

Comparative Analysis of Large Language Models for Bangla Abstractive Text Summarization

Show simple item record

dc.contributor.author Afsar, Md. Farhan
dc.contributor.author Durjoy, Naimur Rahman
dc.date.accessioned 2026-04-12T09:32:57Z
dc.date.available 2026-04-12T09:32:57Z
dc.date.issued 2025-09-16
dc.identifier.uri http://dspace.daffodilvarsity.edu.bd:8080/handle/123456789/16761
dc.description Project Report en_US
dc.description.abstract Abstractive text summarization is a critical challenge in natural language processing (NLP), especially for low-resource languages like Bangla, where data scarcity and weak multilingual adaptation limit progress. This thesis presents a comparative study of three approaches: fine-tuned BanglaT5, fine-tuned mT5, and prompt-engineered GPT. The Bengali Abstractive News Summarization (BANS) dataset was employed, with preprocessing steps such as normalization, tokenization, padding, and truncation to ensure consistency. BanglaT5 and mT5 were fine-tuned using AdamW with crossentropy loss, while GPT was evaluated through zero-shot prompts. Performance was measured with BERTScore and human evaluation by three annotators, who rated outputs on Relevance, Coherence, and Conciseness (1–10 scale). Automatic results show that BanglaT5 achieved the highest BERTScore (F1 0.817% in Bangla embeddings; 0.957% in English embeddings), outperforming mT5 (F1 0.551% in Bangla; 0.765% in English). Human evaluation revealed that GPT consistently scored higher in Relevance 85% and Coherence 84%, while BanglaT5 was rated better for Conciseness 88%, reflecting its ability to produce shorter yet meaningful summaries. These findings highlight the trade-offs between language-specific and general-purpose LLMs: BanglaT5 excels in conciseness and precision, GPT in fluency and relevance, and mT5 underperforms across dimensions. The study concludes that a hybrid approach, combining the precision of BanglaT5 with the fluency of GPT, can significantly advance Bangla summarization and contribute to more inclusive NLP tools for low-resource languages. en_US
dc.description.sponsorship Daffodil International University en_US
dc.language.iso en_US en_US
dc.publisher Daffodil International University en_US
dc.subject Abstractive Text Summarization en_US
dc.subject Bangla Natural Language Processing (NLP) en_US
dc.subject Prompt Engineering en_US
dc.subject BANS Dataset en_US
dc.subject Transfer learning (TL) en_US
dc.subject GPT Models en_US
dc.title Comparative Analysis of Large Language Models for Bangla Abstractive Text Summarization en_US
dc.type Other en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Browse

My Account