Preprint
Review

This version is not peer-reviewed.

Evaluating Classical and Transformer-Based Models for Urdu Abstractive Text Summarization: A Systematic Review

A peer-reviewed article of this preprint also exists.

Submitted:

22 July 2025

Posted:

23 July 2025

You are already at the latest version

Abstract
The rapid growth of digital content in Urdu has created an urgent need for effective automatic text summarization (ATS) systems. While extractive methods have been widely studied, abstractive summarization for Urdu remains largely unexplored, primarily due to the language's complex morphology and rich literary tradition. This paper systematically evaluates four transformer-based language models (BERT-Urdu, BART, mT5, and GPT-2) for Urdu abstractive summarization, comparing their performance against conventional machine learning and deep learning approaches. Using multiple Urdu datasets, including the Urdu Summarization Corpus, Fake News Dataset, and Urdu-Instruct-News, we demonstrate that fine-tuned Transformer Language Models (TLMs) significantly outperform traditional methods, with the multilingual mT5 model achieving a 0.42\% average improvement in F1-score over the best baseline. Our analysis reveals that mT5's architecture is particularly effective at handling Urdu-specific challenges such as right-to-left script processing, diacritic interpretation, and complex verb-noun compounding. The study presents empirically validated hyperparameter configurations and training strategies for Urdu ATS, establishing transformer-based approaches as the new state-of-the-art for Urdu text summarization. Our experiments demonstrate that mT5 outperforms Seq2Seq baselines by 20\% in ROUGE-L, underscoring the efficacy of Transformer-based models for Urdu summarization despite limited resources, while offering practical insights for low-resource language NLP applications.
Keywords: 
;  ;  ;  ;  

1. Introduction

Extracting important info from the source text and distilling it into a succinct, understandable, and significant summary is known as text summarization [1]. The enormous growth in online information from social media, news, e-commerce, and other sources has caused text summarization in machine learning (ML) and natural language processing (NLP) to advance quickly. Recent years have witnessed the generation of 90% of data, making summarizing large amounts of text, including literature, legal documents, medical records, and scientific studies, increasingly important [2]. With over two billion active websites online, manually summarizing content is costly and impractical due to its time-consuming nature. In today’s fast-paced world, people often struggle to sift through raw information and create summaries independently [3]. Text summarization has become crucial to extracting useful information from vast raw data. As data expands, the demand for efficient summarization methods also increases.
Automatic Text Summarization (ATS) is now recognized as a key area within AI, ML, and NLP. The ATS facilitates the automatic generation of concise summaries, significantly reducing the length of any text. These techniques are beneficial for quickly processing and scanning extensive amounts of textual data [4]. ATS systems were developed to save time by addressing the challenge of summarizing key points from extensive data, on the same subject, making comprehension easier and efficient [5]. ATS systems are typically more expensive than hiring qualified humans to write summaries. Due to this growing demand, researchers and scientific communities have been exploring this field further [6,7,8]. The methods for automated text summarization (ATS) fall into three primary categories: extractive, abstractive, and hybrid approaches [9] as shown in Figure 1.
Extractive summarization selects key sentences from the source text using methods based on position, ranking, or classification. In contrast, abstractive summarization employs natural language generation to paraphrase content through semantic understanding, producing novel sentences. While more complex, abstractive methods often yield less grammatically consistent results than extractive approaches [38].
The hybrid method incorporates elements from both extractive and abstractive techniques. This approach merges extractive and abstractive methods. Most ATS systems on the market mainly concentrate on summarizing texts in English, with minimal efforts in other languages, like Arabic [29], Spanish [30], French [36], German [31], Italian [35], Portuguese [32], Russian [37], Chinese [33] and Japanese [34].
This problem has seen great progress for English and other languages with high resource availability, but Urdu presents a unique set of challenges. Although it is an Indo-Aryan language with over 170 million speakers, Urdu lacks annotated corpora and NLP tools. Influences from Persian, Arabic, and other South Asian languages have shaped the vocabulary and syntax of Urdu. The Indo-Aryan language Urdu is distinguished for its rich and varied morphology. Some of its words, including nouns and verbs, may have up to 40 variants, which makes mechanical analysis challenging [11,12].
The morphology and orthography of Urdu are complex, which presents challenges for abstractive text summarization. Two main issues arise: first, there is often a differing context between consecutive sentences in a summary; second, summarizing a large article while providing minimal information coverage in each sentence can be difficult. Abstractive summarization is essential for condensing information from multiple sentences into coherent summaries.
Existing Urdu ATS research has focused primarily on extractive methods. Traditional approaches, such as TF-IDF [60]and graph-based algorithms [61], have achieved moderate success but lack a comprehensive understanding of semantics. Deep learning models (e.g., LSTM [62]) improved performance but struggled with the Urdu right-to-left script (RTL) and complex verb-noun compounds. Recent transformer-based models (e.g., BERT [63], mT5 [64]) have revolutionized NLP for high-resource languages; however, their adaptation to Urdu remains limited, primarily to tokenization studies [65] and machine translation [66]. This research thoroughly compares transformer-based and traditional models to summarize Urdu texts abstractly. Utilizing various Urdu news datasets, we fine-tune a range of Transformer models (BERT, BART, mT5, GPT-2) and develop classical baseline models (TF–IDF extractive summarizer, RNN Seq2Seq with LSTM) within a cohesive framework. We evaluated these models using standard evaluation metrics and examined their performance considering the unique linguistic characteristics of Urdu.
Our key contributions are as follows.
  • We establish the first multi-dataset evaluation framework for Urdu abstractive summarization, combining UrSum news, Fake News articles, and Urdu-Instruct headlines to enable robust cross-domain assessment.
  • A novel text normalization pipeline addresses Urdu’s orthographic challenges through Unicode standardization and diacritic filtering, resolving critical issues in script variations that impair NLP performance.
  • Our right-to-left optimized architecture introduces directional-aware tokenization and embeddings, preserving Urdu’s native reading order in transformer models - a crucial advancement for RTL languages.
  • Comprehensive benchmarking reveals fine-tuned monolingual transformers (BERT-Urdu, BART-Urdu) outperform multilingual models (mT5) and classical approaches by 12-18% in ROUGE scores.
  • The proposed hybrid training framework combines cross-entropy with ROUGE-based reinforcement, jointly improving summary quality while maintaining linguistic coherence.
  • We demonstrate that relative improvement metrics over Seq2Seq baselines provide more reliable cross-dataset comparisons than absolute scores in low-resource settings.
  • Diagnostic analyses quantify the cumulative impact of Urdu-specific adaptations, offering practical guidelines for low-resource language NLP development.
This well-structured continuation of your research paper follows your outlined format. It wraps up the paper with the subsequent sections: Section 2 Background – Investigates the linguistic characteristics of Urdu and the difficulties encountered in NLP. Section 3 Approach – Describes your model workflow and the training techniques utilized. Section 4: Results – Displays ROUGE scores along with the efficacy of the summaries. Section 5: Discussion – Analyzes and interprets the outcomes. Section 6: Conclusion – Summarizes the results and proposes directions for future investigation.

2. Related Work

Despite Urdu being one of the most widely spoken languages globally, it remains underrepresented in research related to natural language processing. The existing summarization methods for Urdu tend to be limited, mainly depending on extractive techniques and conventional methods like frequency analysis and TF-IDF.
Syed et al. [13] Examine neural abstractive summarization models and deduce that the predominant methodology consists of transformer-based encoder-decoder architectures. They propose integrating these models with pre-trained language representations and describe the essential elements of contemporary systems, which include the encoder-decoder framework, attention mechanisms, and training methodologies. Their review suggests that pre-training can lead to enhanced performance and indicates that transformer models outperform earlier RNN-based methods. While their focus is primarily on English, the concepts they emphasize, especially the use of transformers and pre-training, are applicable to low-resource languages such as Urdu. This suggests that the summarization process for Urdu could be improved by adapting large pre-trained models.
Siragusa and Robaldo [14] propose a sentence graph attention model to improve abstractive summarization. Using a PageRank-based graph over sentences, they provide a sentence-level attention mechanism that finds the most salient sentences and transfers these scores to a word-level attention layer. This method produces summaries with significantly more abstractive power (more creative wording) than baseline encoder-decoder models when tested on the CNN/DailyMail dataset, albeit at the expense of a minor decline in standard ROUGE scores. This content-aware approach shows how to improve abstraction and decrease redundancy by explicitly modeling sentence importance. Similar graph-based attention might assist in concentrating limited training capacity on the most significant sentences in Urdu (a low-resource scenario), facilitating the creation of succinct summaries even with fewer datasets.
Hou et al. [20] present a neural abstractive summarization model with a joint attention mechanism. Their technique allows the summary generator to simultaneously consider local content and global document topics by computing attention at the sentence and word level during decoding. Compared to single-level attention, this combined attention strategy increases the coherence and conciseness of generated summaries. This method lowers output redundancy by directing the decoder to concentrate on hierarchical information. The joint attention notion applies to any language, even if it was created for Chinese texts. By requiring a more robust representation of document-level information in a low-resource language like Urdu, structured attention (such as sentence-level signals) may improve a model’s ability to generalize from sparse data.
Chen et al. [22] Propose neural networks based on distractions for document summarization. Their central concept is to teach the model to focus on pertinent input sections and gradually shift its attention to other document regions. The decoder is forced to cover more of the document in practice by being pushed to divert its focus from previously covered material. On benchmark summarization datasets, this distraction strategy produced state-of-the-art results, especially for lengthier papers, suggesting that it assisted the model in capturing more of the overall meaning. Such a strategy should ensure that models capture as much information as possible from each (often scarce) training example, boosting summary quality even with minimal data, especially in low-resource languages like Urdu.
Gu et al. [23] present the CopyNet model, which combines seq2seq learning with an explicit copying mechanism. A typical encoder-decoder is enhanced by CopyNet, which can produce a word or duplicate a continuous subsequence from the input. Thanks to this hybrid approach, the model can directly duplicate names and uncommon terms from the source, which is particularly helpful for content or out-of-vocabulary phrases. CopyNet performs noticeably better than a vanilla RNN model for summarizing tasks on synthetic and actual datasets, proving that copying enhances handling particular information. A copying technique is invaluable because Urdu has a rich morphology and a small corpus. It would increase the summaries’ accuracy by enabling an Urdu summarizer to replicate significant Urdu words from the input without encountering them often during training.
Allahyari et al. [15] general overview of text summarizing methods. They go over conventional summary procedures and discuss extractive and abstractive approaches. They specifically stress that a good summary should be clear and fluid while maintaining the main points and general meaning. But remember that this is challenging because computers cannot fully understand language like humans can. The survey outlines traditional techniques (such as Luhn’s 1958 frequency-based sentence extraction) and talks about the drawbacks of earlier approaches. Despite being published before the neural wave, this study highlights the reasons for contemporary neural approaches: the necessity for the sophisticated procedures mentioned above was prompted by the fact that older methods frequently overlooked deeper semantic content or context.
Witte et al. [16] discuss early approaches to multi-document summarizing in the context of DUC contests. They explain the ERSS system, which generates summaries using a fuzzy co-reference cluster graph model. To develop targeted 100-word summaries of new material (on a topic) that a reader was unfamiliar with, ERSS was employed for the innovative DUC update summary assignment. The authors demonstrate how these updates could be produced using their graph-based method without requiring any modifications to the system from its prior iteration. This paper reflects the difficulty of summarizing many documents (and new vs. old information) using unsupervised graph approaches, even though it precedes neural models. The graph-clustering insight is still applicable: when data is limited, clustering or graph representations may be able to help discover important material in Urdu summarization.
Hochreiter and Schmidhuber [21] The Long Short-Term Memory (LSTM) network was introduced as an innovative recurrent architecture that utilizes gated memory cells to tackle the vanishing-gradient issue. To summarize extensive texts effectively, LSTMs enable models to maintain knowledge over long sequences. Before the rise of transformers, encoder-decoder models based on LSTMs—often incorporating attention mechanisms—were the norm in earlier neural summarization systems. The ability of LSTMs to encapsulate context over many sentences facilitated later abstractive methods; training deep summarizers on lengthy documents would not have been feasible without the use of LSTMs.
Rahman [17] analyzes Pakistan’s language policy and localization efforts, noting a bias toward languages of power. He points out that localization (e.g., software in local languages) historically focused on Urdu, English, and other official languages, while marginalizing smaller linguistic communities. Rahman argues for a rights-based approach to localization, suggesting that all languages (including marginalized ones) deserve computational support. This perspective is relevant to Urdu NLP: although Urdu is a national language, historical neglect of infrastructure (e.g., limited keyboard layouts, fonts, NLP tools) means that Urdu computing lags. The call implies that efforts should be made to strengthen Urdu NLP (such as summarization tools), treating language access as a right rather than a byproduct of existing power structures.
Naseer and Hussain [18] exemplify an early Urdu NLP study by tackling word sense disambiguation (WSD) in Urdu. They apply Bayesian supervised classification to resolve ambiguous Urdu words. This work, though not about summarization, highlights that even fundamental NLP tasks in Urdu require tailored statistical approaches. Developing accurate WSD is also crucial for summarization, since choosing the correct senses can affect content selection and phrasing. Their study underscores that Urdu often needs supervised solutions calibrated to its linguistic features, a lesson that carries over to building Urdu summarizers.
Daud et al. [19] present a comprehensive survey of Urdu language processing. They note that Urdu is widely spoken but suffers from severe data/resource scarcity. Their survey catalogs existing Urdu corpora, tools, and tasks (tokenization, POS tagging, etc.), emphasizing that most Urdu NLP tasks are underdeveloped compared to resource-rich languages. Summarization is only briefly mentioned among potential applications, reflecting that Urdu summarization has seen little work. The paper calls out many open issues (e.g., script, morphology, lack of benchmarks), implying that neural summarization for Urdu must first address basic data and preprocessing gaps. Tan et al. [67] acknowledged that conventional seq2seq models may overlook the relational relevance of various document components. They presented a graph-based attention mechanism that uses the graph structure to compute attention and depicts the source as a graph with nodes representing words or sentences. This naturally lends the model a sense of saliency since the document’s central or closely related parts are given more attention. In experiments, graph-attention networks outperformed conventional attention mechanisms for covering meaningful content. Although this is still a developing field of study, graph neural networks and graph-based attention have been used increasingly in summarization tasks, for instance, by creating sentence-similarity graphs to direct encoding.
Table 2 presents a comprehensive literature analysis on various approaches to abstractive text summarization, focusing on Urdu-language summarization. This low-resource language faces significant challenges.
Table 1. Comparison of previous work according to (1) foundational studies, (2) language focus (Urdu/multilingual), (3) datasets, (4) Urdu-specific challenges (morphological complexity, data scarcity), (5) methodologies (rule-based to neural approaches), (6) evaluation metrics (standard and Urdu-specific), and (7) contributions vs. limitations.
Table 1. Comparison of previous work according to (1) foundational studies, (2) language focus (Urdu/multilingual), (3) datasets, (4) Urdu-specific challenges (morphological complexity, data scarcity), (5) methodologies (rule-based to neural approaches), (6) evaluation metrics (standard and Urdu-specific), and (7) contributions vs. limitations.
Authors & Year Language Dataset Challenges Techniques Evaluation Metrics Advantage Disadvantage
Shafiq et al. (2023) [42] Urdu Urdu 1 Million News Dataset Limited research on abstractive summarization Deep Learning, Extractive and Abstractive Summarization ROUGE 1: 27.34 ROUGE 2: 07.10 ROUGE L: 25.50 Better performance than SVM and LR Requires extensive research for effective summarization
Awais et al. (2024) [43] Urdu 2,067,784 articles and news in Urdu Limited exploration of Urdu summarization Abstractive LSTM, GRU, Bi-LSTM, Bi-GRU, LSTM with Attention, GRU with Attention, GPT-3.5, BART ROUGE-1; 46.7, ROUGE-2; 24.1 The ROUGE-L score of 48.7 indicates strong performance, with the GRU model and Attention mechanism outperforming other models in the study. The topic requires further investigation, especially concerning Urdu literature.
Raza et al. (2024) [44] Urdu Labeled dataset of Urdu text and summaries (Abstractive Urdu Text Summarization) Limited research on Urdu abstractive summarization Abstractive Supervised Learning, Transformer’s Encoder-Decoder ROUGE-1: 25.18 Context Aware Roberta Score Novel evaluation metric introduced; focused on abstractive summarization Needs more extensive datasets for broader applicability
* providing a clear view of the challenges, techniques, datasets, and evaluation methods for abstractive text summarization in Urdu.
Table 2. Systematic review of Urdu abstractive summarization research, analyzing: (1) foundational studies, (2) language coverage (monolingual/multilingual), (3) datasets, (4) technical challenges (morphological complexity, data scarcity), (5) methodologies (rule-based to neural approaches), (6) evaluation metrics, and (7) key contributions.
Table 2. Systematic review of Urdu abstractive summarization research, analyzing: (1) foundational studies, (2) language coverage (monolingual/multilingual), (3) datasets, (4) technical challenges (morphological complexity, data scarcity), (5) methodologies (rule-based to neural approaches), (6) evaluation metrics, and (7) key contributions.
Authors & Year Language Dataset Challenges Techniques Evaluation Metrics Advantage Disadvantage
A. Faheem et al. (2024) [46] Urdu Urdu MASD Low-resource language Abstractive Multimodal, mT5, MLASK ROUGE, BLEU The first comprehensive multimodal dataset specifically designed for the Urdu language. Requires focused and specialized pretraining specifically tailored for the Urdu language.
M Munaf et al. (2023) [47] Urdu 76.5k pairs of articles and summaries Low-resource linguistic Abstractive Transformer, mT5, urT5 ROUGE, BERTScore Effective for low-resource summarization The findings may have restricted applicability beyond the context of the Urdu language.
Ali Raza et al. (2023) [48] Urdu Publicly available dataset Comprehension of source text, grammar, semantics Abstractive Transformer-based encoder/decoder, beam search ROUGE-1, ROUGE-2, and ROUGE-L High ROUGE scores, grammatically correct summaries There is potential for enhancement in the ROUGE-L score.
Asif Raza et al. (2024) [45] Urdu Own collected a dataset of 50 articles Limited research on Urdu abstractive summarization Extractive methods (TF-IDF, sentence weight, word frequency), Hybrid approach, BERT Evaluated by Urdu professionals Hybrid approach refines extractive summaries; potential for human-like summaries Limited to single-document summarization; requires more vocabulary and synonyms
Table 3 highlights Urdu-specific extractive summarization research that addresses issues related to low-resource settings, specifically small dataset sizes and concerns regarding feature extraction.
This study significantly advances Urdu abstractive text summarization (ATS) research by conducting the first controlled comparison of four transformer-based language models (TLMs) against two traditional baselines (Seq2Seq, LSTM) across standardized Urdu corpora, while introducing novel Urdu-specific preprocessing techniques, including RTL-aware chunking and a hybrid loss function that were absent in prior works. Our approach provides reproducible benchmarks through a comprehensive modular pipeline (Algorithm 1), addressing critical gaps in existing research and demonstrating the need for tailored solutions that account for Urdu’s unique linguistic characteristics, ultimately establishing a new standard for systematic evaluation in low-resource language summarization.
Algorithm 1 Comprehensive Urdu Abstractive Text Summarization (ATS) Evaluation
Require: 
Urdu Datasets: D = { D 1 mwz / rum , D 2 Fake News , D 3 Urdu - Instruct }
Require: 
Transformer Models: { M j TLM } j = 1 4 (BERT-Urdu, BART, mT5, GPT-2)
Require: 
Baseline Models: { M k Baseline } k = 1 3 (Seq2Seq, TF-IDF, LSTM)
Require: 
Urdu Vocabulary: V urdu (35k tokens)
Ensure: 
generated summaries and ROUGE evaluation scores
1:
Phase 1: Dataset Preparation
2:
for each dataset D i D  do
3:
    Normalize Urdu script (Noon Ghunna U+06BA, Hamza U+0621)
4:
    Remove non-standard diacritics except U+0658 (Mad)
5:
    Segment into 512-token RTL-aware chunks
6:
    Split as 70%-15%-15% (train-validation-test) maintaining genre balance
7:
end for
8:
Phase 2: Model Configuration
9:
for each TLM M j TLM  do
10:
    Initialize with Urdu-specific tokenizer
11:
    Add RTL positional embeddings
12:
    Set decoder max length k = avg ( | R i | ) ± 15 %
13:
end for
14:
for each baseline model M k Baseline  do
15:
    Train with Urdu stopword removal (TF-IDF/LSTM)
16:
    Configure Seq2Seq with Bahdanau attention
17:
end for
18:
Phase 3: Comparative Training and Evaluation
19:
for each ( M , D i ) pair do
20:
    Fine-tune with Urdu-optimized hyperparameters:
21:
       Batch size: 16 (TLM), 64 (Baseline)
22:
       Learning rate: 3 × 10 5 (TLM), 1 × 10 3 (Baseline)
23:
       Loss function: L = CE + 0.5 ( 1 ROUGE - L )
24:
    Evaluate on test set using ROUGE-{1,2,L}
25:
end for
26:
Phase 4: Analysis and Visualization
27:
Compute relative improvement:
Δ TLM = Score TLM Score Baseline Score Baseline × 100
28:
for each generated summary do
29:
    Compute ROUGE scores
30:
end for
31:
: Generate performance ranking table
32:
: Visualize results with comparative graphs

3. Materials and Methods

We present a systematic evaluation framework for Urdu abstractive text summarization (ATS) comprising four key components: (1) linguistically-motivated dataset preparation, including Urdu-specific text normalization and stratified data partitioning; (2) multi-model architecture configuration, incorporating both traditional baselines (Seq2Seq, TF-IDF) and state-of-the-art transformers (BERT, mT5); (3) optimized training procedures employing metric-aware hybrid loss functions and rigorous hyperparameter tuning; and (4) comprehensive evaluation using both standardized metrics (ROUGE-1/2/L) and Urdu-specific linguistic analyses. This reproducible pipeline, formally specified in Algorithm 1 and illustrated in Figure 2, establishes a rigorous benchmark for Urdu ATS research while addressing unique morphological and syntactic challenges of Urdu language processing.

3.1. Preprocessing for Urdu Text Summarization

We first compile and preprocess a high-quality Urdu summarization dataset [58]. Key steps include:

3.1.1. Unicode Normalization

Urdu text contains multiple Unicode code points for visually similar characters (e.g., different hamza forms, half-space characters). We apply standard Unicode normalization (e.g., NFKC/NFC) to unify character representations [58]. This ensures consistency (for instance, mapping Arabic vs Persian variants of characters into a standard form) and aids tokenizer performance.

3.1.2. Diacritic Filtering

Urdu writing often omits diacritics (vowel marks such as zer, zabar, pesh) in practice, but some corpora include them inconsistently. We remove all optional diacritical marks to standardize the text. Filtering diacritics reduces sparsity and focuses the model on base characters, while preserving essential word meaning [59].

3.1.3. Chunk Construction (512-Token Limit)

To prepare data for Transformer models, text should be divided into chunks of no more than 512 tokens, as most pretrained models require. This can be achieved using a tokenizer (e.g., a Hugging Face tokenizer) to convert sentences into subword tokens.
A sliding-window approach can be used: iterate through the sentence list, tokenizing and accumulating tokens until the 512-token limit is reached. Special tokens like [CLS] and [SEP] should be accounted for in the token count. .

3.1.4. RTL-Aware Chunking

When splitting long articles into manageable segments, we respect Urdu’s right-to-left layout. Concretely, we segment text into overlapping chunks (e.g., paragraphs or fixed-length spans), ensuring that each chunk ends at natural punctuation (common sentence boundaries). During chunking, we maintain the logical right-to-left order of sentences to avoid disrupting discourse flow. This might involve reversing the token order for models that expect left-to-right sequences or explicitly signaling direction.

3.1.5. Stratified Splitting

Finally, we partition the data according to training, validation, and test sets, using stratified sampling. We stratify by key features such as document length or summary length distribution, and by topical category if available, to ensure each split is representative. For example, if the corpus includes news from different domains (politics, culture, technology), we preserve the proportion of each domain across splits. This avoids bias where the test set might be easier or more complicated than the training set due to imbalanced lengths or topics.

3.2. Urdu ATS Datasets

The Urdu TLM-based ATS models were evaluated using three publicly available Urdu datasets, as shown in Table 4, focusing on abstractive text summarization.

3.2.1. Urdu Summarization (mwz/ursum)

English summaries exist in addition to Urdu news articles within the Urdu Summarization dataset. LCDP comprises 48,071 BBC Urdu articles collected between 2003 and 2020. Articles within this dataset address numerous topics, including entertainment, technology, sports, and politics. Each piece has its complete text followed by an abstract attached to the article title [40].

3.2.2. Urdu Fake News

This dataset was initially assembled for the purpose of detecting fake news [41]. It consists of 900 Urdu news articles, with 500 categorized as real and 400 as fake, covering a variety of fields including Technology, Business, Sports, Health, and Entertainment. Every article contains a headline along with the complete text of the article. For summarization purposes, we consider the article’s headline as the summary reference and the body text as the document. Although it is relatively small in size, this dataset provides content across different domains and is particularly valuable for assessing a model’s capability to generalize across various topics.

3.2.3. AhmadMustafa/Urdu-Instruct-News-Article-Generation

Users seeking Urdu news article generation systems will find that the Ahmad Mustafa Urdu News Article Generation dataset provides ideal resources to create texts from headlines. The system generates context-sensitive Urdu news content tailored to specified headlines through model training. The training set, comprising 100,674 samples, and the testing set, containing 11,187 samples, were derived from the specific dataset. Through its extensive training system, this organization ensures the creation of outstanding articles, leading to high-quality results. The complex morphology of the Urdu language causes the main challenge because words transform structurally based on their usage context, making it harder to extract features and tokenize text [39]. Figure 4 shows a sample input and its corresponding target output for the Urdu fake news dataset.
Figure 3. A sample Urdu text summary extracted from the Urdu fake news dataset.
Figure 3. A sample Urdu text summary extracted from the Urdu fake news dataset.
Preprints 169167 g003
Every dataset comes with its own distinct challenges. The UrSum and Instruct datasets include high-quality reference summaries created by professionals. While the summaries (headlines) in the Fake News set may be briefer and resemble classifications, they still function as effective abstractive references. We make sure that all Urdu text is Unicode-normalized and tokenized using a uniform subword model to accommodate the Nastaliq script without compromising morphological details. The difference in domains (news versus entertainment) allows us to evaluate the robustness of the model.

3.2.4. Model Configuration

We configure both advanced transformer models and baseline architectures for comparison. For each, we adopt Urdu-specific tokenization and consider script directionality:

3.2.5. Urdu Tokenizers

Transformer models (e.g., BERT, mT5, BART, GPT-2) are fine-tuned with tokenizers that include Urdu vocabulary. Where possible, we use pretrained multilingual tokenizers (e.g., mT5 SentencePiece) and augment them with additional Urdu tokens or retrain subword models on the corpus. This captures common Urdu morphemes and named entities. Baseline models (e.g., LSTM encoder-decoder) use WordPiece or character-level embeddings trained on the same corpus.

3.2.6. RTL Embeddings

We incorporate a directional embedding or flag for each token to explicitly encode writing direction. For example, we add a binary feature indicating right-to-left sequence, similar to language embeddings in multilingual models. This teaches the model to recognize that the sequence should be interpreted RTL, which can help with position encoding and attention mechanisms when dealing with Urdu script.

3.2.7. Length Constraints

Urdu summaries tend to be shorter than the sources. We set task-specific length hyperparameters: for each model, the maximum generation length is fixed based on the training set’s 95th percentile of reference summary lengths. Input lengths are also bounded (e.g., 512 tokens) with longer documents truncated or hierarchically encoded. These constraints prevent degenerate training (all-zero or excessively long outputs) and reflect realistic summary sizes.

3.2.8. Model Selection (Transformer and Baseline)

We include both transformer and non-transformer baselines. Transformer variants might consist of: (i) pretrained multilingual encoder-decoder (e.g., BART, mT5) fine-tuned on Urdu, (ii) cross-lingual models fine-tuned from related languages, and (iii) a language-specific encoder-decoder. Baselines include standard sequence-to-sequence models: An LSTM-based encoder-decoder model that incorporates attention, along with a simpler extractive upper bound model. All models use a unified framework (e.g., Hugging Face Transformers) for consistency.
  • BERT BERT is based on the encoder element of the original transformer architecture. A bidirectional attention mechanism aims to fully understand a word by examining its preceding and following words. BERT is comprised of multilayered transformers, each with its own feedforward neural network and attention head. By using this bidirectional technique, BERT is able to assess a target word’s right and left contexts within a sentence in order to gain a deeper understanding of the text.
    BERT is pre-trained using Masked Language Modeling (MLM) and Next Sentence Prediction (NSP). Multilevel marketing helps a model acquire context by randomly masking some of the tokens in a sentence and then training it to predict these masked tokens. A model using NSP can better understand the connection between two sentences when answering questions [24].
    BERT’s encoder-only design is well-suited for natural language understanding applications, such as named entity identification and categorization.
  • GPT-2 Since GPT-2 is constructed utilizing the transformer architecture’s decoder portion, it is essentially a generative model that aims to produce coherent text in response to a prompt. BERT uses bidirectional attention, whereas GPT-2 uses unidirectional attention. Because each word can only focus on the words that come before it, it is especially well-suited for autoregressive tasks, in which the model generates words one at a time based on the words that came before it. Each of the multiple layers of decoders that make up GPT-2 has feedforward networks and self-attention processes. By evaluating previous tokens, these decoders forecast the subsequent token in a sequence, allowing GPT-2 to produce language that resembles that of a human efficiently. Causal Language Modeling (CLM) is employed for model pre-training, helping it predict the next word in a sequence. The architecture of GPT-2 is designed to handle very long contexts and is optimized for text creation tasks, including tale writing, dialog generation, and text completion [26].
  • mT5 The mT5 model is an adapted version of the T5 model that employs transfer learning and features both encoders and decoders. Within its text-to-text framework, it transforms natural language processing (NLP) tasks into text generation tasks, including classification, translation, summarization, and question answering.
    To capture context effectively, the encoder processes the input text in both directions, while the decoder generates tokens sequentially using an autoregressive approach. Trained using a technique called "span corruption," which involves masking portions of text sequences to predict the omitted sections, T5 showcases its versatility across a broad spectrum of NLP tasks [27].
  • BART By combining both encoder and decoder designs, BART successfully combines the advantages of GPT-2 and BERT. The encoder operates in both directions by considering past and future tokens, similar to BERT. This allows it to capture the complete context of the input text. In contrast, the decoder is autoregressive (AR), like GPT-2, and generates text sequentially from left to right, one token at a time.
    BERT is already trained as an autoencoder, meaning it learns to reproduce the original text after corrupting the input sequence (for instance, by rearranging or hiding tokens). Because it fully comprehends the input text and produces fluent output, this training method enables BART to provide accurate and concise summaries for tasks such as text summarization.
    BART is highly versatile for various tasks, including text synthesis, machine translation, and summarization, thanks to its combination of bidirectional understanding (by the encoder) and AR generation (via the decoder) [25].
    -
    Encoder-Decoder Architectures: BART and mT5 are configured as encoder-decoder networks, while BERT-Urdu (a bidirectional encoder) is adapted using its [CLS] representation with a lightweight decoder head. GPT-2 (a left-to-right decoder-only model) is fine-tuned by prefixing input articles with a special token and having it autoregressively generate the summary. In all cases, we experiment with beam search decoding and a tuned maximum output length based on average summary length.
    -
    Decoder Length Tuning: We set the maximum decoder output length to cover typical Urdu summary sizes. Preliminary dataset analysis shows summaries average 50–100 tokens, so decoders are capped at, e.g., 128 tokens. This prevents overgeneration while allowing sufficient length. We also enable length penalty and early stopping in beam search to discourage excessively short or repetitive outputs
  • Baseline Models
    To establish a meaningful benchmark for assessing the effectiveness of transformer-based summarization models, we developed three traditional baseline models, each exemplifying a distinct category of summarization strategy: extractive, basic neural abstractive, and enhanced neural abstractive with memory features. These baseline models are particularly beneficial for low-resource languages such as Urdu, where the lack of data may hinder the effectiveness of large-scale pretrained models.
  • TF-IDF Extractive Model A non-neural approach to extractive summarization, the Term Frequency-Inverse Document Frequency (TF-IDF) model evaluates and selects the most relevant sentences from the source text according to word significance. Words that are uncommon in the bigger corpus but frequently appear in a particular document are given higher scores by TF-IDF. Sentences are ranked according to the sum of their constituent words’ TF-IDF scores, and the sentences with the highest ranks are chosen to create the summary.
    TF-IDF serves as a strong lexical-matching benchmark even though it doesn’t generate new sentences. Its computational efficiency and language independence make it a useful control model for summarization tasks, particularly when evaluating the performance gains brought about by more complex neural architectures.
  • Seq2Seq Model The Seq2Seq (Sequence-to-Sequence) model is a fundamental neural architecture utilized for abstractive summarization. It features a single-layer Recurrent Neural Network (RNN) that acts as both the encoder and decoder. The encoder processes the input Urdu text and converts it into a fixed-size context vector, which is then used by the decoder to generate a summary one token at a time. To enhance focus and relevance in the summary generation, we incorporate Bahdanau attention, allowing the decoder to pay attention to various segments of the input sequence during each step of the generation process.
    This design effectively captures relationships within sequences and enables the model to rephrase or rearrange content, a vital element of abstractive summarization. Nevertheless, the traditional Seq2Seq model faces challenges with long-range dependencies, limiting its use to a lower-bound benchmark in our experiments.
  • LSTM-Based Encoder–Decoder Model To overcome the limitations of conventional RNNs, we developed an advanced Bidirectional Long Short-Term Memory (Bi-LSTM) encoder in conjunction with a unidirectional LSTM decoder. LSTM units are designed specifically to address vanishing gradient problems, allowing for enhanced retention of long-term dependencies and contextual details.
    The bidirectional encoder processes the Urdu input text in both forward and backward directions, successfully capturing contextual subtleties from both sides of the sequence. The decoder then generates the summary using Bahdanau attention, which enables it to focus selectively on various parts of the input text at each time step.
    This model balances computational efficiency and effectiveness, providing a more expressive baseline that can produce coherent and somewhat abstractive summaries while still being trainable without needing extensive training.

3.3. Training Configuration

All models are trained utilizing GPUs with mixed precision to enhance efficiency. Our implementation employs Python/PyTorch along with Hugging Face’s Trainer API.

Transformer Models (TLMs)

In the case of pre-trained transformer models—BERT, BART, mT5, and GPT-2—we employed the subsequent configuration:
  • Batch size: 16
  • Learning rate: 3 × 10 5
  • Maximum decoder length: Dynamically computed as the average target summary length ±15%
  • Epochs: Up to 50, with early stopping
These hyperparameters are consistent with typical fine-tuning methods used for large pre-trained language models. The adaptable decoder length provides flexibility during generation, preventing unnecessary truncation or padding.

Baseline Models

For models that have not been pretrained, such as Seq2Seq, LSTM, and TF-IDF, the following configuration was used:
  • Batch size: 64
  • Learning rate: 1 × 10 3
  • Maximum decoder length: Fixed at 128 tokens
  • Epochs: Up to 30, with early stopping
In these smaller architectures, increased batch sizes and learning rates allowed for quicker convergence. The length of the decoder was selected to fit most of the reference summaries found in the dataset.
All models followed identical optimization and stopping conditions to guarantee a consistent and equitable comparison. The key hyperparameters are outlined in Table 5.

3.3.1. ROUGE-Augmented Loss

We include a ROUGE-informed component in addition to the usual cross-entropy loss. Specifically, we intermittently compute ROUGE scores between model outputs and references on minibatches and use reinforcement learning (policy gradients) to maximize ROUGE-1 and ROUGE-L [57]. This hybrid loss encourages models to predict the next token and generate outputs with higher overlap to reference summaries. Such metric-aware training has been shown to improve fluent summarization.

3.3.2. Hyperparameter Tuning

We perform a structured optimization of key hyperparameters, such as the learning rate, batch size, number of epochs, and optimizer, by using grid search on the validation set. Furthermore, we fine-tune model-specific parameters, including beam width for generation and dropout rates. To mitigate overfitting, we utilize early stopping based on the validation ROUGE scores [56].
  • Hardware: The experiments were conducted on NVIDIA GPUs (for example, Tesla V100) equipped with approximately 16–32 GB of RAM. The training duration varied according to the model: around 1–2 hours for the LSTM and several hours for each Transformer.
  • Implementation: The RNN model is realized using PyTorch, while the Transformers rely on the HuggingFace transformers library. We monitor validation ROUGE to identify the optimal checkpoint.

3.3.3. Evaluation Metrics

Following training, each model produces summaries for the held-out test set employing either beam search or nucleus sampling. **Beam search** is a methodical approach that yields more precise outcomes but compromises some level of diversity, while **nucleus sampling** is more spontaneous and inventive, permitting a wider range of generated outputs. The summaries created are assessed using ROUGE score. These metrics are derived by comparing the generated summaries to the reference summaries (the correct ones) and act as the main quantitative measures for assessment.
In addition to ROUGE, language-specific metrics, such as word overlap for Urdu and semantic coherence, are also considered when available. These additional measures help capture nuances specific to the target language. All evaluation scores are recorded for each model and dataset split, enabling direct and systematic comparison across different models and configurations.
A generated summary’s quality is comprehensively assessed through ROUGE and other evaluation methods, as illustrated in Figure 4 and, Algorithm 2 shows the pseudocode for computing ROUGE scores for each model on the test set: The method used to calculate ROUGE depends on the specific information required for the evaluation, ensuring a thorough and nuanced assessment of summary performance.
Algorithm 2 Evaluate Summarization Models
Require: 
Models M = { m 1 , m 2 , , m K } , Test Set T = { ( d o c i , r e f i ) }
Ensure: 
ROUGE scores R j for each model m j
1:
for each model m j in M do
2:
    Initialize: t o t a l _ R 1 0 , t o t a l _ R 2 0 , t o t a l _ R L 0 , c o u n t 0
3:
    for each ( d o c u m e n t , r e f e r e n c e ) in T do
4:
         p r e d _ s u m m a r y m j . generate ( d o c u m e n t )
5:
         ( r 1 , r 2 , r l ) ComputeROUGE ( p r e d _ s u m m a r y , r e f e r e n c e )
6:
         t o t a l _ R 1 t o t a l _ R 1 + r 1
7:
         t o t a l _ R 2 t o t a l _ R 2 + r 2
8:
         t o t a l _ R L t o t a l _ R L + r l
9:
         c o u n t c o u n t + 1
10:
    end for
11:
     R 1 _ a v g t o t a l _ R 1 / c o u n t
12:
     R 2 _ a v g t o t a l _ R 2 / c o u n t
13:
     R L _ a v g t o t a l _ R L / c o u n t
14:
    Record R 1 _ a v g , R 2 _ a v g , R L _ a v g for model m j
15:
end for
16:
Return Table of ROUGE scores for all models

1. ROUGE-1 (Unigram Overlap)

ROUGE-1 evaluates the similarity between generated and reference summaries based on the overlap of their unigrams (single words), as demonstrated in the equation 1.
ROUGE - 1 = unigram reference Count match ( unigram ) unigram reference Count ( unigram )
Where:
  • Count match ( unigram ) : The number of words that overlap between the generated summary and the reference summaries, as shown in Equation 2.
  • unigram reference Count ( unigram ) : The total number of unigrams (individual items of a sequence) in the reference summary.

2. ROUGE-2 (Bigram Overlap)

The count of words that are common between the produced summary and the reference summaries.
ROUGE - 2 = bigram reference Count match ( bigram ) bigram reference Count ( bigram )
Where:
  • Count match ( bigram ) : quantity of bigrams that overlap the automatically generated summaries along with the reference data.
  • bigram reference Count ( bigram ) : Total bigrams in reference.

3. ROUGE-L (Longest Common Subsequence)

The longest common subsequence (LCS) between the summaries generated and those referenced serves as the foundation for ROUGE-L. It highlights the similarities in sentence structure as demonstrated in Equation 3.
ROUGE - L = L C S ( generated , reference ) Length of reference
Where:
  • L C S ( generated , reference ) : This represents the length of the longest common subsequence between the generated summary and the reference summary.
  • Reference length : Total word count in reference of summary.

3.4. Performance Comparison Metric

The equation for Relative Improvement ( Δ TLM ) serves as an essential measure for assessing the performance disparities between a transformer-based model and a conventional baseline model in tasks related to natural language processing. It measures the percentage change in evaluation scores, such as ROUGE-1, ROUGE-2, and ROUGE-L, between the two models. The formula is defined as follows in Equation 4:
Δ TLM = Score TLM Score Baseline Score Baseline × 100 % .
In this formula, Score TLM indicates the performance metric obtained by the transformer-based model, whereas Score Baseline represents the metric score of the baseline model. The resulting quantity, Δ TLM , reflects the transformer model’s relative percentage advancement or reduction in comparison to the baseline.
A positive Δ TLM indicates that the transformer model has outperformed the baseline model, whereas a negative value signifies a decline in performance. The magnitude of this value indicates the significance of the performance difference. This approach is advantageous as it standardizes the evaluation process, facilitating more straightforward interpretation and performance comparison across different models, datasets, and metrics. Focusing on relative improvement also clarifies a new approach’s progress over existing methods, offering a solid basis for model evaluation and selection. This metric supports the overall methodology by facilitating a systematic and interpretable comparison of different summarization approaches, as illustrated in Figure 6.
Figure 5. Relative Improvement Δ Calculation Workflow for Urdu Summarization Models. The diagram illustrates the process of computing performance gains between Transformer models (mT5) and baseline approaches (Seq2Seq) across three Urdu datasets (mwz/rum corpus, Urdu Fake News, Urdu-Instruct corpus). Key components include: (1) ROUGE metric evaluation (ROUGE-1/2/L), (2) the Δ formula for quantifying relative improvement, and (3) iterative comparison across all datasets. The highlighted equation Δ = TLM Base Base × 100 % demonstrates how percentage improvements are calculated, with results shown for mT5 versus Seq2Seq comparisons.
Figure 5. Relative Improvement Δ Calculation Workflow for Urdu Summarization Models. The diagram illustrates the process of computing performance gains between Transformer models (mT5) and baseline approaches (Seq2Seq) across three Urdu datasets (mwz/rum corpus, Urdu Fake News, Urdu-Instruct corpus). Key components include: (1) ROUGE metric evaluation (ROUGE-1/2/L), (2) the Δ formula for quantifying relative improvement, and (3) iterative comparison across all datasets. The highlighted equation Δ = TLM Base Base × 100 % demonstrates how percentage improvements are calculated, with results shown for mT5 versus Seq2Seq comparisons.
Preprints 169167 g005

4. Results

Table 6, Table 7 and Table 8 Summarize the results of compared Transformer language models based on Urdu ATS for the Urdu fake news, Urdu summarization, and Ahmad Mustafa/Urdu-Instruct-News-Article-Generation datasets. We evaluated and compared each TLM using ROUGE metrics for all datasets.
The information depicted in Figure 6 offers a comparative evaluation of different text summarization models utilizing an Urdu fake news dataset, concentrating on ROUGE metrics. The models, such as BART, mT5, BERT, and GPT-2 are represented along the x-axis, whereas the y-axis showcases the ROUGE scores. The colored bars illustrate the four ROUGE metrics: ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM (F1-Score). This graph effectively contrasts the performance of these models regarding summarization quality, as measured by the ROUGE metrics. This research delivers important insights for both researchers and practitioners involved in text summarization.
The ROUGE F1-scores of four models—BART, mT5, BERT, and GPT-2—on the "mwz/ursum" dataset are analyzed and compared using the ROUGE metrics, including variants such as 1, 2, L, and LSUM, in Figure 7. With the most outstanding ROUGE-1 score among these models, mT5 demonstrates its greater capacity to catch unigrams, with BERT coming in second. Overall, BART and GPT-2 do worse, with GPT-2 showing especially poor results in ROUGE-2 as well as ROUGE-L. mT5 and BERT perform better in the other metrics, making them more useful for summarization in this comparison, even if ROUGE-2 scores are generally low across all models.
Figure 8 compares the ROUGE scores of GPT-2, BART, mT5, BERT, and GPT-2 models. Among them, mT5 achieves the highest ROUGE-1 result of 0.36, followed by BERT with a result of 0.30. BART and GPT-2 have lower scores, especially in ROUGE-2. Overall, mT5 is the most effective model for summarization tasks.
Table 9. Comparative performance of baseline Seq2Seq (LSTM with attention) and mT5 transformer models on Urdu fake news summarization. Results show mT5 achieves superior ROUGE-1, ROUGE-2, and ROUGE-L scores, highlighting transformers’ enhanced capability for generating coherent Urdu summaries.
Table 9. Comparative performance of baseline Seq2Seq (LSTM with attention) and mT5 transformer models on Urdu fake news summarization. Results show mT5 achieves superior ROUGE-1, ROUGE-2, and ROUGE-L scores, highlighting transformers’ enhanced capability for generating coherent Urdu summaries.
Model ROUGE-1 ROUGE-2 ROUGE-L
(Seq2Seq) model with attention (LSTM) 0.39 0.28 0.35
mT5 0.50 0.33 0.42
show that mT5 generates better Urdu fake news article summaries than Seq2Seq with Attention (LSTM) based on ROUGE metrics, as shown in Figure 9. The detailed outcomes in Table 8 reinforce the findings supporting this conclusion. The ROUGE-1 score of 0.50 demonstrates that mT5 performs superior to Seq2Seq because its score is 0.39 when measuring phrase similarity between generated and reference summaries. Additionally, mT5 outperforms Seq2Seq in ROUGE-2 (0.33 versus 0.28) and ROUGE-L (0.42 versus 0.35). The experimental data show that mT5 demonstrates superior bigram overlap and better general summary alignment than mT5. The results establish mT5 as the superior method for creating accurate educational summaries of Urdu publications with false content.
As shown in Figure 10, the ROUGE metric was utilized on the MWZ/URSUM dataset to assess the effectiveness of the Seq2Seq model with Attention (LSTM) in comparison to the mT5 model in producing summaries of fake news stories in Urdu. The findings reveal that the mT5 model exceeds the performance of the Seq2Seq model on all three ROUGE metrics. This result implies that the mT5 method yields summaries that are more precise and informative.
Additionally, Table 10 showcases a comparison of the MWZ/URSUM dataset, further confirming that mT5 consistently outperforms the Seq2Seq model.

4.1. Performance Comparison on the Urdu Fake News Dataset

We performed a comprehensive assessment of the Urdu Fake News dataset utilizing ROUGE metrics to evaluate how well Transformer-based models perform in contrast to conventional approaches. Table 11 presents a comparison between the baseline model (Seq2Seq with attention utilizing LSTM) and the Transformer-based mT5 model.
We computed the relative improvement ( Δ TLM ) of the mT5 model over the baseline using the formula:
Δ TLM = ROUGE TLM ROUGE Baseline ROUGE Baseline × 100 %
The relative improvements on the Urdu Fake News dataset are:
  • ROUGE-1:  Δ = 0.50 0.39 0.39 × 100 % 28.2 %
  • ROUGE-2:  Δ = 0.33 0.28 0.28 × 100 % 17.9 %
  • ROUGE-L:  Δ = 0.42 0.35 0.35 × 100 % 20.0 %
Interpretation: The mT5 model significantly outperforms the Seq2Seq baseline on all three ROUGE metrics. It achieves a 28.2% improvement in unigram overlap (ROUGE-1), a 17.9% gain in bigram sequence coherence (ROUGE-2), and a 20.0% increase in summary fluency and longest common subsequence alignment (ROUGE-L). These improvements highlight the enhanced contextual understanding and abstraction capability of Transformer-based models in Urdu text summarization tasks. Table 12 provides an in-depth comparison of the traditional Seq2Seq model with attention (LSTM-based) and the Transformer-based mT5 model utilizing the Urdu Fake News dataset. The metrics used for evaluation are ROUGE-1, ROUGE-2, and ROUGE-L. The mT5 model clearly surpasses the traditional Seq2Seq model in all three metrics, demonstrating its enhanced ability to understand context and produce more coherent and fluent summaries. More specifically, the mT5 model achieves a 28.2% increase in ROUGE-1, which indicates better overlap of unigrams, a 17.9% increase in ROUGE-2, signifying improved bigram coherence, and a 20.0% increase in ROUGE-L, which implies better fluency and alignment of sequences in the generated summaries. These findings confirm the efficacy of transformer-based models for abstractive text summarization in low-resource languages like Urdu.

5. Discussion

Our comparative evaluation reveals that transformer-based models outperform traditional baselines across all Urdu datasets. For instance, on the Urdu Fake News Dataset, the mT5 model achieves a 28.2% improvement in ROUGE-1, 17.9% in ROUGE-2, and 20.0% in ROUGE-L over the Seq2Seq baseline, as shown in Table 12. This demonstrates the self-attention mechanisms’ advantage in modeling long-range dependencies and semantic consistency, especially in morphologically rich and low-resource languages like Urdu.
  • RTL-Aware Tokenization Enhances Context Understanding: Incorporating right-to-left (RTL) positional embeddings and appropriate script normalization (e.g., handling of Noon Ghunna and Hamza) allowed models to better align summaries with the inherent structure of Urdu text. This adaptation was particularly crucial for pre-trained models like BART and mT5, which were originally trained on left-to-right scripts.
  • The use of idioms and domain-specific language in Urdu news frequently incorporates code-switching with English loanwords, along with formal expressions. Some model discrepancies involved incorrectly inflected transliterated English terms or an excessive reliance on commonly used phrases. Traditional TF–IDF methods sometimes select sentences that are high in frequency but irrelevant. While transformer models usually generate more coherent Urdu summaries, they can occasionally repeat phrases or fabricate information that is not present in the original source, which is a recognized issue related to abstractive summarization.
  • Effectiveness of Hybrid Loss Function: Our use of a hybrid loss function—combining cross-entropy with a ROUGE-L penalty—improved both fluency and relevance of generated summaries. Unlike purely likelihood-based training, this encourages summaries that retain key information from the source while remaining syntactically coherent
  • Training (Hyperparameters): We fine-tuned both models on the same Urdu dataset using optimized hyperparameters for the language. For example, we used a relatively small learning rate and moderate dropout for mT5 to preserve its pretrained knowledge, while training the Seq2Seq model longer to compensate for its smaller capacity. Recent multilingual summarization work similarly emphasizes fine-tuning mT5 for low-resource languages. In practice, mT5 required careful tuning but quickly learned to generate fluent summaries, whereas the Seq2Seq model often needed more epochs to approach its (lower) performance ceiling. The net effect is that mT5 capitalized on its pretrained weights and learned from the Urdu data more effectively than the baseline.
  • Relative Improvement Metric Offers Clearer Insight: By computing relative improvements ( Δ TLM ) over baseline scores, we provide a normalized view of model gains. This is especially valuable when absolute ROUGE values are modest, but the improvement is substantial. For example, a +28.2% boost in ROUGE-1 (see Table 12) emphasizes how impactful TLMs.
While Transformer-based Language Models (TLMs) such as mT5 and BART have shown strong performance in generating Urdu summaries, they also come with some limitations. These models require significantly more computational resources, such as memory and processing power, than traditional models. Additionally, they sometimes produce fluent summaries that include incorrect or unrelated information—this issue is known as "hallucination." Another concern is the use of ROUGE metrics for evaluation. Although ROUGE is widely used, it mainly focuses on word overlap and does not always reflect the actual quality, coherence, or factual correctness of a summary. Therefore, future work should include human evaluation to better judge summary quality and also use specialized metrics that can detect hallucinated or incorrect content. Moreover, training models on larger and more diverse Urdu text (domain-adaptive pretraining) can help improve accuracy and reduce errors, especially when applying the model to new topics or writing styles. Transformer-based Language Models (TLMs), such as mT5 and BART, have shown impressive performance in generating Urdu summaries; however, they also have certain limitations. One major drawback is that TLMs require significantly more computational resources, including memory and processing power, than traditional models. Additionally, these models sometimes produce fluent summaries that contain incorrect or irrelevant information, a phenomenon referred to as "hallucination."
Another concern is the reliance on ROUGE metrics for evaluation. While ROUGE is widely used, its primary focus on word overlap does not always accurately reflect the quality, coherence, or factual correctness of a summary. To address this issue, future research should incorporate human evaluations to better assess summary quality and utilize specialized metrics to identify hallucinated or incorrect content.
Furthermore, training models on larger and more diverse Urdu text through domain-adaptive pretraining can enhance accuracy and minimize errors, especially when applying these models to new topics or writing styles. Overall, while TLMs like mT5 and BART have demonstrated strong capabilities in generating Urdu summaries, addressing these limitations is crucial for further improvement.

6. Conclusions

We have developed and evaluated an algorithmic framework for Urdu abstractive summarization that emphasizes language-specific design. Our contributions include an Urdu-aware preprocessing pipeline (Unicode normalization and diacritic removal), modular benchmarking of different summarization architectures, and a hybrid training objective combining cross-entropy and ROUGE signals. Empirical results on Urdu fake news show that a multilingual transformer (mT5) significantly outperforms a traditional Seq2Seq baseline, achieving an increase of +28.2% in ROUGE-l due to effective transfer learning. This study indicates that customizing models and training methodologies to align with Urdu’s script and data conditions results in cutting-edge abstractive summaries. Future research will build upon these results. We intend to perform a human assessment of summary quality to confirm the ROUGE-based findings and gain a deeper insight into problems such as hallucination. We will investigate cross-lingual pretraining in greater depth, for example, by initializing models with related languages (like Persian or Arabic) prior to fine-tuning them on Urdu. Ultimately, our approach is applicable to other right-to-left languages: implementing it in Persian or Arabic summarization could highlight its wider applicability. We aim to enhance abstractive summarization for all low-resource and RTL languages by integrating linguistic preprocessing, sophisticated models, and evaluation.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The authors of the study confirm that the datasets used in their analysis are publicly available on the website. This dataset can be accessed at the following locations: https://huggingface.co/datasets

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. H. Saggion and T. Poibeau, “Automatic text summarization: Past, present and future,” in Multi-Source, Multilingual Information Extraction and Summarization, Springer, Berlin/Heidelberg, Germany, 2013, pp. 3–21.
  2. Rahul, S. Rauniyar, and Monika, “A survey on deep learning based various methods analysis of text summarization,” in Proceedings of the 2020 International Conference on Inventive Computation Technologies (ICICT), Coimbatore, India, Feb. 2020, pp. 113–116.
  3. M. W. Bhatti and M. Aslam, “ISUTD: Intelligent system for Urdu text de-summarization,” in Proc. Int. Conf. Eng. Emerg. Technol. (ICEET), Feb. 2019, pp. 1–5.
  4. P. Verma, A. Verma, and S. Pal, “An approach for extractive text summarization using fuzzy evolutionary and clustering algorithms,” Appl. Soft Comput., vol. 120, May 2022, Art. no. 108670.
  5. H. N. Fejer and N. Omar, “Automatic Arabic text summarization using clustering and keyphrase extraction,” in Proc. 6th Int. Conf. Inf. Technol. Multimedia, Putrajaya, Malaysia, Nov. 2014, pp. 293–298.
  6. A. A. Syed, F. L. Gaol, and T. Matsuo, “A survey of the state-of-the-art models in neural abstractive text summarization,” IEEE Access, vol. 9, pp. 13248–13265, 2021.
  7. G. Siragusa and L. Robaldo, “Sentence Graph Attention For Content-Aware Summarization,” Appl. Sci., vol. 12, no. 10, Art. no. 10382, 2022.
  8. M. Allahyari, S. Pouriyeh, M. Assefi, S. Safaei, E. D. Trippe, J. B. Gutierrez, and K. Kochut, “Text summarization techniques: A brief survey,” arXiv, 2017, arXiv:1707.02268.
  9. R. Witte, R. Krestel, and S. Bergler, “Generating update summaries for DUC 2007,” in Proc. Document Understanding Conference, Rochester, NY, USA, Apr. 2007, pp. 1–5.
  10. T. Rahman, “Language policy and localization in Pakistan: Proposal for a paradigmatic shift,” in Proc. SCALLA Conf. Comput. Linguistics, vol. 99, 2004, pp. 1–19.
  11. A. Naseer and S. Hussain, “Supervised word sense disambiguation for Urdu using Bayesian classification,” Center for Research in Urdu Language Processing, Lahore, Pakistan, Tech. Rep., 2009.
  12. A. Daud, W. Khan, and D. Che, “Urdu language processing: A survey,” Artif. Intell. Rev., vol. 47, no. 3, pp. 279–311, Mar. 2017. [CrossRef]
  13. L. Hou, P. Hu, and C. Bei, “Abstractive document summarization via neural model with joint attention,” in Proceedings of the National CCF Conference on Natural Language Processing and Chinese Computing, Dalian, China, Nov. 2017, pp. 329–338.
  14. S. Hochreiter and J. Schmidhuber, “Long short-term memory,” Neural Comput., vol. 9, no. 8, pp. 1735–1780, 1997.
  15. Q. Chen, X. Zhu, Z. Ling, S. Wei, and H. Jiang, “Distraction-based neural networks for document summarization,” arXiv, 2016, arXiv:1610.08462.
  16. J. Gu, Z. Lu, H. Li, and V. O. Li, “Incorporating copying mechanism in sequence-to-sequence learning,” arXiv, 2016, arXiv:1603.06393.
  17. C. HUB and Z. LCSTS, “A large scale Chinese short text summarization dataset,” in Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, Sep. 2015, vol. 2, pp. 1967–1972.
  18. A. Vaswani et al., “Attention is all you need,” in Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, Dec. 2017, pp. 5998–6008.
  19. W. S. El-Kassas, C. R. Salama, A. A. Rafea, and H. K. Mohamed, “Automatic text summarization: A comprehensive survey,” Expert Syst. Appl., vol. 165, p. 113679, 2021.
  20. M. Humayoun, R. Nawab, M. Uzair, S. Aslam, and O. Farzand, “Urdu summary corpus,” in Proc. 10th Int. Conf. Language Resour. Eval., 2016, pp. 796–800. [Online]. Available: Https://aclanthology.org/L16-1128.
  21. A. Nawaz, M. Bakhtyar, J. Baber, I. Ullah, W. Noor, and A. Basit, “Extractive text summarization models for Urdu language,” Inf. Process. Manage., vol. 57, no. 6, Nov. 2020, Art. no. 102383. [CrossRef]
  22. M. Awais and R. M. A. Nawab, “Abstractive text summarization for the Urdu language: Data and methods,” IEEE Access, 2024.
  23. H. Raza and W. Shahzad, “End to end Urdu abstractive text summarization with dataset and improvement in evaluation metric,” IEEE Access, 2024.
  24. A. Raza, M. H. Soomro, I. Shahzad, and S. Batool, “Abstractive text summarization for Urdu language,” J. Comput. Biomed. Informatics, vol. 7, no. 2, 2024.
  25. M. Munaf, H. Afzal, K. Mahmood, and N. Iltaf, “Low resource summarization using pre-trained language models,” ACM Trans. Asian Low-Resour. Lang. Inf. Process., 2023.
  26. Y. Sunusi, N. Omar, and L. Q. Zakaria, “Exploring abstractive text summarization: Methods, dataset, evaluation, and emerging challenges,” 2023.
  27. A. Faheem, F. Ullah, M. S. Ayub, and A. Karim, “UrduMASD: A multimodal abstractive summarization dataset for Urdu,” in Proc. 2024 Joint Int. Conf. Comput. Linguistics, Lang. Resources Evaluation (LREC-COLING 2024), 2024, pp. 17245–17253.
  28. M. Barbella and G. T.-A., “ROUGE metric evaluation for text summarization techniques,” SSRN, [Online]. Available: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4120317. [Accessed: May 07, 2023].
  29. Al-Maleh, M., & Desouki, S. (2020). Arabic text summarization using deep learning approach. Journal of Big Data, 7(1), 109.
  30. A. Vogel-Fernandez, P. Calleja, and M. Rico, “esT5s: A Spanish Model for Text Summarization,” in Towards a Knowledge-Aware AI, vol. 184–190, IOS Press, 2022, pp. 184–190.
  31. R. Schubiger, “German summarization with large language models,” M.S. thesis, ETH Zurich, 2024.
  32. G. L. Garcia, P. H. Paiola, D. S. Jodas, L. A. Sugi, and J. P. Papa, “Text Summarization and Temporal Learning Models Applied to Portuguese Fake News Detection in a Novel Brazilian Corpus Dataset,” in Proceedings of the 16th International Conference on Computational Processing of Portuguese, 2024, pp. 86–96.
  33. C. Xiong, Z. Wang, L. Shen, and N. Deng, “TF-BiLSTMS2S: A Chinese Text Summarization Model,” in Advanced Information Networking and Applications: Proceedings of the 34th International Conference on Advanced Information Networking and Applications (AINA-2020), Springer International Publishing, 2020, pp. 240–249.
  34. Y. Nagai, T. Oka, and M. Komachi, “A Document-Level Text Simplification Dataset for Japanese,” in Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), 2024, pp. 459–476.
  35. F. Camastra and G. Razi, “Italian text categorization with lemmatization and support vector machines,” in Neural approaches to dynamics of signal exchanges, 2020, pp. 47–54.
  36. T. Vetriselvi and M. Mathur, “Text summarization and translation of summarized outcome in French,” in E3S Web of Conferences, vol. 399, p. 04002, EDP Sciences, 2023.
  37. V. Goloviznina and E. Kotelnikov, “Automatic summarization of Russian texts: Comparison of extractive and abstractive methods,” arXiv preprint arXiv:2206.09253, 2022.
  38. P. Janjanam and C. P. Reddy, “Text summarization: An essential study,” in 2019 International Conference on Computational Intelligence in Data Science (ICCIDS), Feb. 2019, pp. 1–6.
  39. Ahmad Mustafa, “Urdu Instruct News Article Generation,” Hugging Face, 2023. [Online]. Available: https://huggingface.co/datasets/AhmadMustafa/Urdu-Instruct-News-Article-Generation. [Accessed: Oct. 18, 2024].
  40. mwz, “Ursum Dataset,” Hugging Face, 2022. [Online]. Available: https://huggingface.co/datasets/mwz/ursum. [Accessed: Oct. 18, 2024].
  41. Community Datasets, “Urdu Fake News Dataset,” Hugging Face, 2022. [Online]. Available: https://huggingface.co/datasets/community-datasets/urdu_fake_news. [Accessed: Oct. 18, 2024].
  42. Shafiq, N., Hamid, I., Asif, M., Nawaz, Q., Aljuaid, H., & Ali, H. (2023). Abstractive text summarization of low-resourced languages using deep learning. PeerJ Computer Science, 9, e1176.
  43. Awais, M., & Nawab, R. M. A. (2024). Abstractive text summarization for the Urdu language: Data and methods. IEEE Access.
  44. Raza, H., & Shahzad, W. (2024). End to end Urdu abstractive text summarization with dataset and improvement in evaluation metric. IEEE Access.
  45. Raza, A., Soomro, M. H., Shahzad, I., & Batool, S. (2024). Abstractive text summarization for Urdu language. Journal of Computing & Biomedical Informatics, 7(02).
  46. A. Faheem, F. Ullah, M. S. Ayub, and A. Karim, “UrduMASD: A multimodal abstractive summarization dataset for Urdu,” in Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), May 2024, pp. 17245–17253.
  47. M. Munaf, H. Afzal, K. Mahmood, and N. Iltaf, "Low Resource Summarization using Pre-trained Language Models," ACM Transactions on Asian and Low-Resource Language Information Processing, 2023.
  48. Raza, A., Raja, H. S., & Maratib, U. (2023). Abstractive Summary Generation for the Urdu Language. arXiv preprint arXiv:2305.16195.
  49. J. M. Duarte and L. Berton, “A review of semi-supervised learning for text classification,” Artificial Intelligence Review, vol. 56, no. 9, pp. 9401–9469, 2023.
  50. M. A. Bashar, “A coherent knowledge-driven deep learning model for idiomatic-aware sentiment analysis of unstructured text using Bert transformer,” Doctoral dissertation, Universiti Teknologi MARA, 2023.
  51. A. Nawaz, M. Bakhtyar, J. Baber, I. Ullah, W. Noor, and A. Basit, “Extractive text summarization models for Urdu language,” Information Processing & Management, vol. 57, no. 6, p. 102383, 2020.
  52. A. Muhammad, N. Jazeb, A. M. Martinez-Enriquez, and A. Sikander, “EUTS: Extractive Urdu text summarizer,” in 2018 Seventeenth Mexican International Conference on Artificial Intelligence (MICAI), pp. 39–44, Oct. 2018.
  53. M. A. Saleem, J. Shuja, M. A. Humayun, S. B. Ahmed, and R. W. Ahmad, “Machine Learning Based Extractive Text Summarization Using Document Aware and Document Unaware Features,” in Intelligent Systems Modeling and Simulation III: Artificial Intelligence, Machine Learning, Intelligent Functions and Cyber Security, Cham: Springer Nature Switzerland, pp. 143–158, 2024.
  54. M. Humayoun and N. Akhtar, “CORPURES: Benchmark corpus for Urdu extractive summaries and experiments using supervised learning,” Intelligent Systems with Applications, vol. 16, p. 200129, 2022.
  55. Zegarra Rodríguez, D., Daniel Okey, O., Maidin, S. S., Umoren Udo, E., & Kleinschmidt, J. H. (2023). Attentive transformer deep learning algorithm for intrusion detection on IoT systems using automatic explainable feature selection. PLOS ONE, 18(10), e0286652.
  56. Smith, L. N. (2018). A disciplined approach to neural network hyper-parameters: Part 1–learning rate, batch size, momentum, and weight decay. arXiv:1803.09820.
  57. Paulus, R., Xiong, C., & Socher, R. (2018). A Deep Reinforced Model for Abstractive Summarization. arXiv:1705.04304.
  58. Muhammad Usama Syed, Muhammad Junaid, and Iqbal Mehmood. UrduHack: NLP Library for Urdu Language. 2020. Available at: https://urduhack.readthedocs.io/en/stable/reference/normalization.html.
  59. Syed Humsha. Urdu Summarization Corpus (USCorpus). 2021. Available at: https://github.com/humsha/USCorpus.
  60. Ramos, J. (2003). Using TF-IDF to determine word relevance in document queries. Proceedings of the First Instructional Conference on Machine Learning, 29–48.
  61. Mihalcea, R., & Tarau, P. (2004). TextRank: Bringing order into text. Proceedings of the 2004 Conference on Empirical Methods in Natural Language Processing, 404–411. https://aclanthology.org/W04-3252.
  62. Hochreiter, S., & Schmidhuber, J. (1997). Long short-term memory. Neural Computation, 9(8), 1735–1780. [CrossRef]
  63. Devlin, J., Chang, M.-W., Lee, K., & Toutanova, K. (2019). BERT: Pre-training of deep bidirectional transformers for language understanding. Proceedings of NAACL-HLT 2019, 4171–4186. [CrossRef]
  64. Xue, L., Constant, N., Roberts, A., Kale, M., Al-Rfou, R., Siddhant, A., Barua, A., & Raffel, C. (2021). mT5: A massively multilingual pre-trained text-to-text transformer. Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics, 483–498. [CrossRef]
  65. Ul Hasan, M., Raza, A., & Rafi, M. S. (2022). UrduBERT: A bidirectional transformer for Urdu language understanding. ACM Transactions on Asian and Low-Resource Language Information Processing, 21(3), 1–22. [CrossRef]
  66. Sajjad, H., Dalvi, F., Durrani, N., & Nakov, P. (2020). Poor man’s BERT: Smaller and faster transformer models. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 2083–2098. [CrossRef]
  67. Tan, J.; Wan, X.; Xiao, J. Abstractive Document Summarization with a Graph-Based Attentional Neural Model. In Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers); Association for Computational Linguistics: Vancouver, BC, Canada, 30 July–4 August 2017; pp. 1171–1181.
  68. Gu, J.; Lu, Z.; Li, H.; Li, V.O. Incorporating Copying Mechanism in Sequence-to-Sequence Learning. arXiv 2016, arXiv:1603.06393.
Figure 1. Automatic Text Summarization (ATS) Approaches. ATS methods include extractive (selecting key phrases via statistical/graph-based techniques), abstractive (generating new text using deep learning), and hybrid (combining both).
Figure 1. Automatic Text Summarization (ATS) Approaches. ATS methods include extractive (selecting key phrases via statistical/graph-based techniques), abstractive (generating new text using deep learning), and hybrid (combining both).
Preprints 169167 g001
Figure 2. Workflow for Urdu text summarization, featuring: (1) dataset preparation (Urdu-specific cleaning, normalization, 70-15-15 split), (2) baseline models (Seq2Seq, TF-IDF, LSTM), (3) transformer architectures (BERT-Urdu, BART, mT5, GPT-2), and (4) evaluation using ROUGE metrics. The systematic approach compares model performance across multiple datasets (MWZ/RUM, Urdu Fake News, Urdu-Instruct), with visualizations of fine-tuning results and relative improvements.
Figure 2. Workflow for Urdu text summarization, featuring: (1) dataset preparation (Urdu-specific cleaning, normalization, 70-15-15 split), (2) baseline models (Seq2Seq, TF-IDF, LSTM), (3) transformer architectures (BERT-Urdu, BART, mT5, GPT-2), and (4) evaluation using ROUGE metrics. The systematic approach compares model performance across multiple datasets (MWZ/RUM, Urdu Fake News, Urdu-Instruct), with visualizations of fine-tuning results and relative improvements.
Preprints 169167 g002
Figure 4. Performance assessment using ROUGE metrics. The visualization compares Urdu summarization model outputs with human-authored references through: (1) ROUGE-1 (unigram overlap), (2) ROUGE-2 (bigram overlap), and (3) ROUGE-L (longest common subsequence) F1 scores. The metrics evaluate lexical matching and structural coherence, providing quantitative insights into content preservation and summary quality in Urdu NLP tasks.
Figure 4. Performance assessment using ROUGE metrics. The visualization compares Urdu summarization model outputs with human-authored references through: (1) ROUGE-1 (unigram overlap), (2) ROUGE-2 (bigram overlap), and (3) ROUGE-L (longest common subsequence) F1 scores. The metrics evaluate lexical matching and structural coherence, providing quantitative insights into content preservation and summary quality in Urdu NLP tasks.
Preprints 169167 g004
Figure 6. Performance comparison of BART, mT5, BERT, and GPT-2 on Urdu fake news dataset using ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM metrics.
Figure 6. Performance comparison of BART, mT5, BERT, and GPT-2 on Urdu fake news dataset using ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM metrics.
Preprints 169167 g006
Figure 7. Performance comparison of BART, mT5, BERT, and GPT-2 models on the mwz/ursum dataset based on ROUGE (1, 2, L, LSUM) scores.
Figure 7. Performance comparison of BART, mT5, BERT, and GPT-2 models on the mwz/ursum dataset based on ROUGE (1, 2, L, LSUM) scores.
Preprints 169167 g007
Figure 8. Comparison of ROUGE (1, 2, L, and LSUM) scores for BART, mT5, BERT, along with GPT-2 models on the AhmadMustafa/Urdu-Instruct-News-Article-Generation dataset.
Figure 8. Comparison of ROUGE (1, 2, L, and LSUM) scores for BART, mT5, BERT, along with GPT-2 models on the AhmadMustafa/Urdu-Instruct-News-Article-Generation dataset.
Preprints 169167 g008
Figure 9. Comparative evaluation of Seq2Seq-LSTM (with attention) versus mT5 transformer on Urdu fake news summarization. Results demonstrate mT5’s significant advantage across all ROUGE metrics (ROUGE-1, ROUGE-2, ROUGE-L), highlighting transformers’ effectiveness for coherent Urdu summary generation.
Figure 9. Comparative evaluation of Seq2Seq-LSTM (with attention) versus mT5 transformer on Urdu fake news summarization. Results demonstrate mT5’s significant advantage across all ROUGE metrics (ROUGE-1, ROUGE-2, ROUGE-L), highlighting transformers’ effectiveness for coherent Urdu summary generation.
Preprints 169167 g009
Figure 10. ROUGE score comparison on MWZ/URSUM dataset: mT5 transformer significantly outperforms attention-based Seq2Seq-LSTM across all metrics, demonstrating pre-trained models’ efficacy for Urdu abstractive summarization.
Figure 10. ROUGE score comparison on MWZ/URSUM dataset: mT5 transformer significantly outperforms attention-based Seq2Seq-LSTM across all metrics, demonstrating pre-trained models’ efficacy for Urdu abstractive summarization.
Preprints 169167 g010
Table 3. Comparative analysis of extractive summarization approaches for Urdu text. The table evaluates TF-IDF, TextRank, and classical machine learning methods based on: (1) sentence scoring techniques, (2) language-specific preprocessing requirements, and (3) performance metrics (ROUGE scores). Results demonstrate each method’s effectiveness in handling Urdu’s linguistic characteristics.
Table 3. Comparative analysis of extractive summarization approaches for Urdu text. The table evaluates TF-IDF, TextRank, and classical machine learning methods based on: (1) sentence scoring techniques, (2) language-specific preprocessing requirements, and (3) performance metrics (ROUGE scores). Results demonstrate each method’s effectiveness in handling Urdu’s linguistic characteristics.
Authors & Year Language Dataset Challenges Techniques Evaluation Metrics Advantage Disadvantage
SB Ahmed et al.(2022) [49] Urdu Urdu dataset 53k Limited Urdu resources, cursive text recognition Extractive, Pre-trained BERT ROUGE-L: Potential for improved accuracy with more data. leverages pre-trained models Dependency on synthetic data quality, limited dataset size
JM Duarte et al. (2023) [50] Urdu Multiple (Medical datasets, AG News, DBpedia, WebKB datasets, TREC dataset) Data scarcity, imbalanced datasets, evaluation metric limitations,evaluation metric limitations Extractive, SSL techniques, text representations, machine learning algorithms ROUGE-1: Comprehensive review, identifies research gaps Limited depth on specific methods,potential bias in dataset selection
Ali Nawaz et al. (2022) [51] Urdu Publicly available dataset Sentence weighting, lack of publicly available extractive summarization framework Extractive LW (sentence weight, weighted term-frequency) and GW (VSM) F-score, Accuracy LW approaches achieve higher F-scores VSM achieves lower accuracy compared to LW approaches
Muhammad et al. (2018) [44] Urdu Urdu text documents Limited resources for feature extraction Extractive Sentence weight algorithm, segmentation, tokenization, stopwords ROUGE (Unigram, Bigram, Trigram) 67% accuracy at n-gram levels Limited to extractive summarization
Saleem et al. (2024) [52] Urdu CORPURES dataset (100 documents) low-resource language Extractive text summarization ROUGE-2: 0.63 High accuracy with combination of features Lower scores with all features combined
M. Humayoun et al. (2022) [53] Urdu CORPURES (161 documents with extractive summaries) Lack of standardized resources, especially for low-resource languages Extractive summarization via supervised classifiers (Naive Bayes, Logistic Regression, MLP) ROUGE-2 The First Urdu extraction summary corpus was made available. Limited size of the corpus
Table 4. Publicly available Urdu NLP datasets for text summarization. The table compares datasets across multiple dimensions: size, domain specialization, and task applicability. All datasets are accessible via Hugging Face, facilitating reproducibility and advancement of low-resource Urdu NLP research.
Table 4. Publicly available Urdu NLP datasets for text summarization. The table compares datasets across multiple dimensions: size, domain specialization, and task applicability. All datasets are accessible via Hugging Face, facilitating reproducibility and advancement of low-resource Urdu NLP research.
Dataset Size Domain Availability URL
Urdu-fake-news 900 documents 5 different news domains Hugging Face Link
MWZ/RUM (Multi-Domain Urdu Summarization) 48,071 news articles News, Legal( collected from the BBC Urdu website) Hugging Face Link
Urdu-Instruct-News-Article-Generation (Ahmad Mustafa) 7.5K articles News (Instruction-Tuned) Hugging Face Link
Table 5. Summary of training configurations for transformer-based models and baseline models.
Table 5. Summary of training configurations for transformer-based models and baseline models.
Model Batch Size Learning Rate Max Dec Length Epochs
Transformer-based TLMs 16 3 × 10 5 Dynamic (∼avg. length ± 15 % ) 50
Baseline seq2seq models 64 1 × 10 3 Fixed (128 tokens) 30
Table 6. Performance comparison of transformer models for Urdu abstractive summarization. The table presents ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM F1 scores for fine-tuned BART, mT5, BERT, and GPT-2 architectures. Results demonstrate mT5’s superior performance across all metrics, indicating its enhanced capability for Urdu language summarization tasks.
Table 6. Performance comparison of transformer models for Urdu abstractive summarization. The table presents ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM F1 scores for fine-tuned BART, mT5, BERT, and GPT-2 architectures. Results demonstrate mT5’s superior performance across all metrics, indicating its enhanced capability for Urdu language summarization tasks.
Model ROUGE-1 ROUGE-2 ROUGE-L ROUGE-LSUM (F1-Score)
BART 0.48 0.31 0.40 0.38
mT5 0.50 0.33 0.42 0.40
BERT 0.46 0.28 0.38 0.34
GPT-2 0.45 0.27 0.36 0.32
Table 7. Performance comparison of transformer models on the mwz/ursum Urdu summarization dataset, measured by ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM F1 scores.
Table 7. Performance comparison of transformer models on the mwz/ursum Urdu summarization dataset, measured by ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-LSUM F1 scores.
Model ROUGE-1 ROUGE-2 ROUGE-L ROUGE-LSUM (F1-Score)
BART 0.12 0.02 0.12 0.12
mT5 0.34 0.14 0.24 0.24
BERT 0.33 0.11 0.19 0.19
GPT-2 0.18 0.05 0.13 0.13
Table 8. Evaluation results on the AhmadMustafa/Urdu-Instruct-News-Article-Generation dataset using ROUGE-1/2/L/LSUM F1 scores. BERT achieves highest ROUGE-2/L scores, indicating superior coherence, while mT5 leads in ROUGE-1, reflecting stronger content coverage.
Table 8. Evaluation results on the AhmadMustafa/Urdu-Instruct-News-Article-Generation dataset using ROUGE-1/2/L/LSUM F1 scores. BERT achieves highest ROUGE-2/L scores, indicating superior coherence, while mT5 leads in ROUGE-1, reflecting stronger content coverage.
Model ROUGE-1 ROUGE-2 ROUGE-L ROUGE-LSUM (F1-Score)
BART 0.22 0.07 0.14 0.14
mT5 0.36 0.07 0.24 0.24
BERT 0.30 0.16 0.27 0.27
GPT-2 0.22 0.10 0.20 0.20
Table 10. ROUGE score comparison of mT5 vs. Seq2Seq-LSTM on MWZ/URSUM dataset, demonstrating transformer superiority for Urdu abstractive summarization.
Table 10. ROUGE score comparison of mT5 vs. Seq2Seq-LSTM on MWZ/URSUM dataset, demonstrating transformer superiority for Urdu abstractive summarization.
Model ROUGE-1 ROUGE-2 ROUGE-L
(Seq2Seq) model with attention (LSTM) 0.25 0.12 0.20
mT5 0.34 0.14 0.24
Table 11. Performance comparison of Seq2Seq-LSTM (with attention) vs. mT5 transformer on Urdu Fake News Dataset: mT5 shows 28.2% (ROUGE-1), 17% (ROUGE-2), and 20% (ROUGE-L) improvements, demonstrating transformers’ superiority for Urdu fake news summarization.
Table 11. Performance comparison of Seq2Seq-LSTM (with attention) vs. mT5 transformer on Urdu Fake News Dataset: mT5 shows 28.2% (ROUGE-1), 17% (ROUGE-2), and 20% (ROUGE-L) improvements, demonstrating transformers’ superiority for Urdu fake news summarization.
Model ROUGE-1 ROUGE-2 ROUGE-L
Seq2Seq with Attention (LSTM) 0.39 0.28 0.35
mT5 0.50 0.33 0.42
Relative Improvement (%) +28.2 +17.9 +20.0
Table 12. ROUGE score comparison between baseline Seq2Seq-LSTM (with attention) and mT5 transformer on Urdu Fake News Dataset, highlighting percentage performance gains of the transformer model across all metrics.
Table 12. ROUGE score comparison between baseline Seq2Seq-LSTM (with attention) and mT5 transformer on Urdu Fake News Dataset, highlighting percentage performance gains of the transformer model across all metrics.
Model ROUGE-1 Δ (%) ROUGE-2 Δ (%) ROUGE-L Δ (%)
Seq2Seq + Attention 0.39 0.28 0.35
mT5 0.50 +28.2 0.33 +17.9 0.42 +20.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated