Preprint
Article

This version is not peer-reviewed.

An Efficient Hybrid Deep Learning Approach for Detecting Online Abusive Language

Submitted:

13 March 2026

Posted:

16 March 2026

You are already at the latest version

Abstract
The digital age has expanded social media and online forums, allowing free expression for nearly 45% of the global population. Yet, it has also fueled online harassment, bullying, and harmful behaviors like hate speech and toxic comments across social networks, messaging apps, and gaming communities. Studies show 65% of parents notice hostile online behavior, and one-third of adolescents in mobile games experience bullying. A substantial volume of abusive content is generated and shared daily, not only on the surface web but also within dark web forums. Creators of abusive comments often employ specific words or coded phrases to evade detection and conceal their intentions. To address these challenges, we propose a hybrid deep learning model that integrates BERT, CNN, and LSTM architectures with a ReLU activation function to detect abusive language across multiple online platforms, including YouTube comments, online forum discussions, and dark web posts. The model demonstrates strong performance on a diverse and imbalanced dataset containing 77,620 abusive and 272,214 non-abusive text samples (ratio 1:3.5), achieving approximately 99% across evaluation metrics such as Precision, Recall, Accuracy, F1-score, and AUC. This approach effectively captures semantic, contextual, and sequential patterns in text, enabling robust detection of abusive content even in highly skewed datasets, as encountered in real-world scenarios.
Keywords: 
;  ;  ;  ;  ;  

1. Introduction

The rise of the digital age and the Internet has fueled widespread use of platforms like social media and forums, enabling free expression. Nearly 45% of the global population uses social media, often becoming addicted. However, this has also led to harassment, bullying, and harmful behavior, such as hate speech, toxic comments, and sharing obscene content. Digital technologies make such behavior possible across various online platforms, including social media, messaging, and gaming sites [1]. Social media has become a major platform for hostile behavior, with 65% of parents worldwide acknowledging its prevalence. Additionally, one-third of adolescents who play mobile games have reported being victims of bullying [2]. According to [3], 44% of internet users in the United States have personally encountered various forms of online harassment, with 28% experiencing severe hostility.
Automatically detecting and analyzing online abuse text presents significant challenges due to the complexity of language, contextual ambiguity, the dynamic evolution of terminology, and the sheer volume of data. These challenges are further amplified in the detection of Abuse Material shared on the dark web, where privacy and anonymity are prioritized, making it difficult to trace perpetrators [4]. Additionally, offenders often employ sophisticated evasion techniques, such as the use of code words, slang, cryptic abbreviations, and other forms of linguistic obfuscation, to bypass detection mechanisms and conceal illicit activities [5]. Furthermore, legal and ethical constraints necessitate careful handling of sensitive data, limiting the extent to which automated systems can operate effectively. These factors collectively make the development of robust AI-driven detection methods both critical and highly challenging [6].
Machine learning (ML) and deep learning (DL) techniques have been widely adopted to develop abusive text detection models, including Support Vector Machines (SVM) [7], Random Forest [8], Decision Trees [9], Naïve Bayes (NB) [10], K-Nearest Neighbors (KNN) [11], Long Short-Term Memory (LSTM) [12], Bi-directional LSTM (Bi-LSTM) [13], Convolutional Neural Networks (CNN) [14], and Bidirectional Encoder Representations from Transformers (BERT) [15]. However, each individual ML/DL technique has limitations, such as sensitivity to noise, difficulty handling large-scale data, or an inability to capture complex relationships. Hybrid models that combine ML/DL techniques can leverage the strengths of multiple algorithms to address these weaknesses [16]. Therefore, we propose a hybrid DL model combining BERT, CNN, LSTM, and ReLU activation. This integration improves prediction accuracy, especially for complex or imbalanced datasets, enhances generalization, reduces overfitting, and adapts to a wider range of tasks. These benefits are particularly valuable in domains like abusive text detection, where language can be highly variable and nuanced.
The contributions of our research can be articulated as follows:
  • Building an integrated dataset of 77,620 abusive and 272,214 non-abusive text samples from three sources to create a diverse, imbalanced collection that enables comprehensive model evaluation across real-world scenarios.
  • Proposing a hybrid DL model that effectively captures abusive text in both English and Romanized scripts by integrating BERT, CNN, LSTM, and the ReLU activation function. The model achieves very high performance with a Precision of 0.991, Recall of 0.986, Accuracy of 0.995, F1-score of 0.989, and AUC of 0.992, as evaluated using five-fold cross-validation.
  • Conducting a comprehensive comparison of the proposed model with traditional ML and standalone DL baselines using a diverse benchmark dataset comprising YouTube comments, forum discussions, and dark web posts, evaluated through 5-fold cross-validation across multiple performance metrics.
The structure of this paper is as follows. In Section 2, we review relevant research. Section 3 outlines our methodology, providing our hybrid DL model and its comprising modules in detail. In Section 4, we describe the dataset and evaluate both our model and selected baseline models using various metrics, followed by a detailed performance analysis. Finally, Section 5 concludes the paper and offers insights for future research directions.

3. Methodology

In this paper, we propose a hybrid DL model for detecting online abusive comments. The model architecture, illustrated in Figure 1, is designed to predict the abuse polarity of a text and classify it accordingly. It consists of three main modules: Preprocessing, Feature Extraction and Classification. In the Preprocessing module, BERT is employed as the word embedding model to generate feature vectors. The Feature Extraction module integrates CNN and LSTM models to capture both local and sequential features of the text. Finally, the Classification module applies a ReLU activation function to produce the final output.
The purpose of this module combination is to enhance detection accuracy compared to single-model approaches when applied to complex data from diverse sources, although it requires longer computation time. The hybrid architecture leverages the complementary strengths of BERT, CNN, and LSTM: BERT captures the full contextual meaning of words through bidirectional analysis, CNN effectively extracts salient textual features, and LSTM retains past information via its cell states, enhancing the model’s ability to learn sequential dependencies.

3.1. Word Embedding Layer

The initial phase of our model involves transforming the input text into a sequence of word embeddings—dense vector representations that encode semantic relationships among words. In this work, we deployed a pre-trained BERT model tailored for the Online Abusive Comments (OAC) dataset. After fine-tuning its parameters, BERT serves as a feature extractor, generating contextualized embeddings for the proposed hybrid framework. The OAC dataset is passed through the BERT model to generate feature vectors, which are then fed into the subsequent layers of the hybrid architecture. The BERT-base-uncased architecture was used to produce contextualized embeddings with a hidden size of 768.

3.2. Convolutional Neural Networks Model

A Convolutional Neural Network (CNN) is a form of feedforward neural network that processes data sequentially from input to output, without any feedback loops. It employs a deep layered architecture [29], usually starting with convolutional and pooling (or subsampling) layers that extract and transform features from the input data. These are then followed by one or more fully connected layers that perform the final classification.
While the general architecture of CNNs is independent of the dimensionality of the input data, their implementation depends on it. The dimensionality determines how subsampling filters slide across the data. In natural language processing, a one-dimensional convolutional layer (1D CNN) with m filters is commonly applied, producing an m-dimensional feature vector for each document n-gram. These feature vectors are then aggregated using max-pooling, followed by a ReLU (Rectified Linear Unit) activation function. The resulting output is then forwarded to a fully connected (linear) layer, which performs the ultimate classification step, as illustrated in Figure 2. To capture local contextual patterns in BERT embeddings, a sequence of three 1D convolutional layers with 512, 256, and 128 filters, respectively, and a kernel size of 3 is applied.
Let w i : n R d denote the input text consisting of n words, where each word is represented as a d-dimensional embedding vector. The sequence of embeddings forms a d × n matrix that serves as the input to the convolutional layer, where filters are applied across the text.
For each l-word n-gram, we define:
c i = [ w i , , w i + l 1 ] R ( d × l ) , 0 i n l
For each filter f j R d × l , the convolution operation is computed as the inner product c i , f j . The resulting convolutional feature maps are collected into a matrix F R n × m , where m is the number of filters.
A max-pooling operation is then applied across the n-gram dimension:
p j = max i ( F i j )
The pooled features are passed through a ReLU non-linearity to introduce activation.
Finally, the extracted feature representation is passed through a fully connected layer that produces a probability distribution across the target classes. The class corresponding to the highest probability is then chosen as the final prediction.

3.3. Long Short-Term Memory Model

The Long Short-Term Memory (LSTM) network is a specialized variant of the Recurrent Neural Network architecture [30]. Each LSTM unit comprises three main gates—the forget gate, input gate, and output gate—as well as input/output components and a memory cell that allows the model to learn and retain long-term dependencies. The structure of an LSTM block is illustrated in Figure 3. In our model, the extracted features are passed through an LSTM layer with a hidden size of 500 to effectively capture sequential dependencies.
The forget gate regulates the extent to which information from the previous cell state c t 1 is preserved:
f t = σ W f · h t 1 , x t + b f
The input gate determines the amount of new information to incorporate into the cell state, while the candidate cell state c ˜ t is calculated as follows:
i t = σ ( W i [ h t 1 , x t ] + b i ) , c ˜ t = tanh ( W c [ h t 1 , x t ] + b c )
The updated cell state is then:
c t = f t · c t 1 + i t · c ˜ t
Finally, the output gate controls the proportion of the cell state that is exposed to the hidden output:
o t = σ ( W o [ h t 1 , x t ] + b o ) , h t = o t · tanh ( c t )
Each gate has its own learnable weights and biases, enabling the network to determine how much past information to retain, how much new input to incorporate, and how much of the internal state to expose at each time step.

3.4. Fully Connected Layer and Output

The context vector produced by the LSTM layer with an attention mechanism is fed into a fully connected (dense) layer utilizing a ReLU activation function. This layer enhances the high-level features learned by the CNN and LSTM modules, enabling the model to capture more intricate and non-linear relationships within the data.
Next, the output from the dense layer is passed to the final output layer, which employs an appropriate activation function (e.g., softmax) to generate the abuse classification probabilities. The model is optimized using the categorical cross-entropy loss function, and the class label with the highest probability is selected as the final prediction.

4. Model Evaluation

4.1. Our Dataset

To evaluate model performance, our dataset includes 77,620 abusive and 272,214 non-abusive text samples, integrated from three datasets. This diverse and imbalanced collection reflects real-world conditions, where abusive content is considerably less frequent than normal content. It enhances the model’s generalization across different contexts, supports robust evaluation under realistic class distributions, and enables fair comparison with other imbalanced benchmark datasets. The combined datasets include:
1.
Darkweb Dataset: We created the Darkweb dataset comprising 4,600 samples extracted from 352,000 dark web forum posts collected in 2022. It includes 2,500 child sexual abuse (CSA)-related and 2,100 non-CSA samples. Among them, 2,000 CSA and 100 non-CSA samples contain at least one sexual abuse phrase, see more details in [5]. Figure 4 presents blurred word clouds depicting single words and two-word phrases associated with sexual abuse, which were extracted from the post contents.
2.
PAN12 Dataset1: Developed for the CLEF 2012 competition, this dataset aims to detect predatory behavior in online chats. It includes 198,054 conversations, consisting of 4,029 abusive and 194,025 non-abusive samples.
3.
Roman Urdu Dataset: Contains 147,180 YouTube comments, evenly divided into abusive (73,590) and non-abusive (73,590) classes [31].

4.2. Experimental Design and Metrics

To assess the model’s performance, we use accuracy (ACC), precision (P), recall (R), and F1-score (F1), which are computed from the confusion matrix values: true positives (TP), false positives (FP), true negatives (TN), and false negatives (FN). These metrics are defined as follows: accuracy is computed by A C C = T P + T N T P + F P + T N + F N , precision by P = T P T P + F P , recall by R = T P T P + F N , and the F1-score by F 1 = 2 · P · R P + R [32].
The AUC (Area Under the Curve) metric is also used to assess a model’s overall classification performance. It represents the area beneath the ROC curve, which graphs the True Positive Rate against the False Positive Rate at various threshold settings. Higher AUC values, approaching 1, signify greater classification accuracy and model effectiveness.
The experiments were carried out in a Kaggle Notebook environment featuring an NVIDIA Tesla T4 GPU with 16 GB of memory, of which up to 15 GB was available per session, alongside an Intel(R) Xeon(R) CPU @ 2.00 GHz with four logical cores. The setup used Python 3.10, with the PyTorch 2.5 framework and Hugging Face’s Transformers library for implementing and fine-tuning the BERT-based model. The environment also provided up to 30 GB of RAM and 57.6 GB of temporary storage, which proved sufficient for training, validating, and evaluating the model’s performance efficiently. We use 5-fold cross-validation during model development where the reported classification performance and execution times correspond to the averaged scores across the five runs.

4.3. Results and Discussion

The average Precision, F1-score, and AUC of the NB, LR, SVM, CNN, and LSTM models using TF-IDF ( t f . i d f ) and Word2Vec ( w 2 v ) feature representations are presented in Figure 5, averaged across five-fold cross-validation. Overall, models leveraging Word2Vec embeddings generally outperform their TF-IDF counterparts across most evaluation metrics, reflecting the benefits of semantic representations in capturing contextual meaning. The only exception is the NB model, where NB t f . i d f achieves higher Precision (0.696) and F1-score (0.783) than NB w 2 v , although the latter records a slightly higher AUC (0.906 vs. 0.892), indicating a marginally better discriminative ability.
Among the remaining models, LR w 2 v , SVM w 2 v , CNN w 2 v , and LSTM w 2 v consistently achieve superior results. Specifically, LR w 2 v outperforms LR t f i d f in Precision (0.897 vs. 0.867), F1-score (0.864 vs. 0.787), and Accuracy (0.903 vs. 0.845). SVM w 2 v also shows improvements over SVM t f i d f in F1-score (0.914 vs. 0.884) and AUC (0.937 vs. 0.847). CNN w 2 v also surpasses CNN t f i d f across all three evaluation metrics (Precision: 0.819 vs. 0.693, F1-score: 0.759 vs. 0.442, AUC: 0.829 vs. 0.610), demonstrating the advantage of semantic embeddings. Similarly, LSTM w 2 v achieves the best overall performance among all models, with a Precision of 0.94, F1-score of 0.922, and AUC of 0.935. These results highlight the effectiveness of DL and kernel-based models when combined with Word2Vec embeddings. Therefore, the best-performing models in this comparison are NB t f . i d f , LR w 2 v , SVM w 2 v , CNN w 2 v , and LSTM w 2 v .
Figure 6 presents and conpares the confusion matrix components across our proposed hybrid model, BERT and the the best-performing ML/DL models based on the t f . i d f or w 2 v . The proposed hybrid model achieved the highest number of TP (15,307) and TN (54,305), indicating its superior ability to correctly identify both abusive and non-abusive comments. It also recorded the lowest FP (138) and FN (217), demonstrating a strong balance between precision and recall. In contrast, traditional models such as NB and LR showed weaker performance, with significantly higher FP and FN. For example, NB t f . i d f produced 6,070 FP and 1,619 FN, while CNN w 2 v also struggled with 4,546 FN. BERT and LSTM w 2 v performed considerably better, reducing misclassifications and achieving stronger overall accuracy; however, our model consistently outperformed them in all four metrics.
Table 1 summarizes the experimental results across all models. The proposed hybrid model achieved the best overall performance, with the highest Precision (0.991), Recall (0.986), Accuracy (0.995), F1-score (0.989), and AUC (0.992). This indicates its strong ability to correctly identify both abusive and non-abusive comments with minimal errors. BERT and LSTM w 2 v also performed competitively, achieving high precision and recall but slightly lower accuracy and F1-score compared to the hybrid model. Traditional ML models, including NB t f . i d f , LR w 2 v , and SVM w 2 v , delivered moderate performance, with NB showing the lowest precision (0.696) and CNN w 2 v yielding the weakest overall results. In terms of computational efficiency, NB t f . i d f and LR w 2 v had the fastest training and prediction times, while DL models, particularly BERT and the hybrid model, required substantially longer processing times due to their complexity. Overall, the hybrid model offers the best trade-off between classification performance and reliability, albeit at the cost of increased computational time. These results highlight the effectiveness of integrating DL architectures in a hybrid framework to enhance both sensitivity and specificity in online abusive comment detection.
Additionally, the Hybrid model also achieves the highest performance on the ROC curve, as shown in Figure 7, with an averaged AUC of 99.2%. The curve almost touches the top-left corner, indicating an optimal balance between the true positive rate and the false positive rate.

5. Conclusions and Future Work

This paper presents an efficient hybrid deep learning model that combines BERT, CNN, and LSTM architectures with a ReLU activation function to detect abusive language across multiple online platforms. By effectively integrating these components, the model captures semantic, contextual, and sequential dependencies in text, enabling robust classification of abusive and non-abusive content. Evaluation on a large, diverse, and imbalanced dataset—including YouTube comments, online forum posts, and dark web content—demonstrated its effectiveness under realistic conditions.
Experimental results showed that the proposed hybrid model outperformed traditional ML and standalone DL methods. It achieved strong results with a Precision of 0.991, Recall of 0.986, Accuracy of 0.995, F1-score of 0.989, and AUC of 0.992, confirming its high discriminative ability and robustness. Although computationally more demanding than simpler models, this trade-off remains acceptable for real-world applications requiring reliable performance. Overall, the hybrid model provides an effective and flexible approach to abuse detection, serving as a strong foundation for future research on content moderation and digital safety.
Future work will focus on extending this framework to multilingual and code-mixed text, improving interpretability through XAI methods, applying Ontology [33,34], exploiting Knowledge Graph [35,36] and incorporating multimodal [37] features such as images or user metadata. Additionally, optimizing computational efficiency via model compression or pruning will support real-time deployment in large-scale social media monitoring systems [38].

Acknowledgments

This publication has emanated from research conducted with the financial support of Taighde Éireann – Research Ireland under Grant number 12/RC/2289_P2.

References

  1. Chinivar, S.; et al. Online offensive behaviour in socialmedia: Detection approaches, comprehensive review and future directions. Entertainment Computing 2023, 45, 100544. [Google Scholar] [CrossRef]
  2. Zuckerman, A. 60 Cyberbullying Statistics: 2020/2021 Data, Insights & Predictions. Accessed: 2025-04-02.
  3. Petrosyan, A. Impact of Online Hate and Harassment in the U.S. 2020, 2025. Accessed: 2025-09-02.
  4. Ngo, V.M.; Thorpe, C.; McKeever, S. Analysing Child Sexual Abuse Activities in the Dark Web based on an Efficient CSAM Detection Algorithm. In Proceedings of the The 2nd Annual Trust and Safety Research Conference, Stanford, USA, September 2023. [Google Scholar]
  5. Ngo, V.M.; Mckeever, S.; Thorpe, C. Identifying Online Child Sexual Texts in Dark Web through Machine Learning and Deep Learning Algorithms. In Proceedings of the the APWG.EU Technical Summit and Researchers Sync-Up (APWG.EU-Tech 2023). CEUR Workshop Proceedings, 2024; pp. 1–6. [Google Scholar]
  6. Ngo, V.M.; Thorpe, C.; Dang, C.N.; Mckeever, S. Investigation, Detection and Prevention of Online Child Sexual Abuse Materials: A Comprehensive Survey. In Proceedings of the 2022 RIVF Int. Conf. on Computing and Communication Technologies (RIVF), 2022; pp. 707–713. [Google Scholar]
  7. Muneer, A.; Fati, S.M. A comparative analysis of machine learning techniques for cyberbullying detection on twitter. Future Internet 2020, 12. [Google Scholar] [CrossRef]
  8. Amrit, C.; Paauw, T.; Aly, R.; Lavric, M. Identifying child abuse through text mining and machine learning. Expert Systems with Applications 2017, 88, 402–418. [Google Scholar] [CrossRef]
  9. Talpur, B.A.; O’Sullivan, D. Cyberbullying severity detection: A machine learning approach. PLOS ONE 2020, 15, 1–19. [Google Scholar] [CrossRef]
  10. Mckeever, S.; Thorpe, C.; Ngo, V.M. Determining Child Sexual Abuse Posts based on Artificial Intelligence. In Proceedings of the 2023 International Society for the Prevention of Child Abuse & Neglect Congress (ISPCAN-2023), Edinburgh, UK, 2023; pp. 24–27. [Google Scholar]
  11. Wadud, M.A.H.; et al. How can we manage Offensive Text in Social Media - A Text Classification Approach using LSTM-BOOST. International Journal of Information Management Data Insights 2022, 2, 100095. [Google Scholar] [CrossRef]
  12. Chadaga, K.; et al. An Explainable Framework to Predict Child Sexual Abuse Awareness in People Using Supervised Machine Learning Models. Journal of Technology in Behavioral Science 2023, 1–17. [Google Scholar] [CrossRef]
  13. Li, T.; Zeng, Z.; Sun, S. A two-stage cyberbullying detection based on multi-view features and decision fusion strategy. Applied Intelligence 2025, 55, 294. [Google Scholar] [CrossRef]
  14. Marshan, A.; Nizar, F.N.M.; Ioannou, A.; Spanaki, K. Comparing Machine Learning and Deep Learning Techniques for Text Analytics: Detecting the Severity of Hate Comments Online. Information Systems Frontiers 2023. [Google Scholar] [CrossRef]
  15. Ngo, V.M.; Gajula, R.; Thorpe, C.; Mckeever, S. Discovering child sexual abuse material creators’ behaviors and preferences on the dark web. Child Abuse & Neglect 2024, 147, 106558. [Google Scholar]
  16. Dang, C.N.; Moreno-García, M.N.; De la Prieta, F. Hybrid deep learning models for sentiment analysis. Complexity 2021, 2021, 9986920. [Google Scholar] [CrossRef]
  17. Mazzarello, O.; Gagné, M.E.; Langevin, R. Risk factors for sexual revictimization and dating violence in young adults with a history of child sexual abuse. Journal of Child & Adolescent Trauma 2022, 15, 1113–1125. [Google Scholar] [CrossRef]
  18. Owusu-Addo, E.; et al. Prevalence and determinants of sexual abuse among adolescent girls during the COVID-19 lockdown and school closures in Ghana: A mixed method study. Child Abuse & Neglect 2023, 135, 105997. [Google Scholar]
  19. Aguerri, J.C.; Molnar, L.; Miró-Llinares, F. Old crimes reported in new bottles: the disclosure of child sexual abuse on Twitter through the case# MeTooInceste. Social Network Analysis and Mining 2023, 13, 27. [Google Scholar] [CrossRef]
  20. Gangwar, A.; González-Castro, V.; Alegre, E.; Fidalgo, E. AttM-CNN: Attention and metric learning based CNN for pornography, age and Child Sexual Abuse (CSA) Detection in images. Neurocomputing 2021, 445, 81–104. [Google Scholar] [CrossRef]
  21. Kissos, L.; Goldner, L.; Butman, M.; Eliyahu, N.; Lev-Wiesel, R. Can artificial intelligence achieve human-level performance? A pilot study of childhood sexual abuse detection in self-figure drawings. Child Abuse & Neglect 2020, 109, 104755. [Google Scholar]
  22. Laranjeira, C.; Macedo, J.; Avila, S.; Santos, J. Seeing without Looking: Analysis Pipeline for Child Sexual Abuse Datasets. In Proceedings of the Proc. of 2022 ACM Conf. on Fairness, Accountability, and Transparency (FAccT’22). ACM, 2022; pp. 2189–2205. [Google Scholar]
  23. Alam, I.; Basit, A.; Ziar, R.A. Utilizing Age-Adaptive Deep Learning Approaches for Detecting Inappropriate Video Content. Human Behavior and Emerging Technologies 2024, 2024, 7004031. [Google Scholar] [CrossRef]
  24. Hole, M.; Frank, R.; Logos, K.; Westlake, B.; Michalski, D.; Bright, D.; Afana, E.; Brewer, R.; Ross, A.; Swearingen, T.; et al. Developing automated methods to detect and match face and voice biometrics in child sexual abuse videos. Trends and Issues in Crime and Criminal Justice 2022, 1–15. [Google Scholar]
  25. Cook, D.; Zilka, M.; DeSandre, H.; Giles, S.; Maskell, S. Protecting Children from Online Exploitation: Can a Trained Model Detect Harmful Communication Strategies? Proceedings of the Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society. ACM 2023, AIES ’23, 5–14. [Google Scholar]
  26. Puentes, J.; et al. Guarding the Guardians: Automated Analysis of Online Child Sexual Abuse. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), 2023; pp. 3730–3734. [Google Scholar]
  27. Kaur, S.; Singh, S.; Kaushal, S. Deep learning-based approaches for abusive content detection and classification for multi-class online user-generated data. International Journal of Cognitive Computing in Engineering 2024, 5, 104–122. [Google Scholar] [CrossRef]
  28. Mosa, M.A. Optimizing text classification accuracy: a hybrid strategy incorporating enhanced NSGA-II and XGBoost techniques for feature selection. Progress in Artificial Intelligence 2025. [Google Scholar] [CrossRef]
  29. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: an overview and application in radiology. Insights into imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [PubMed]
  30. Palangi, H.; et al. Deep sentence embedding using long short-term memory networks: Analysis and application to information retrieval. IEEE/ACM transactions on audio, speech, and language processing 2016, 24, 694–707. [Google Scholar] [CrossRef]
  31. Akhter, M.P.; Jiangbin, Z.; Naqvi, I.R.; Abdelmajeed, M.; Sadiq, M.T. Automatic detection of offensive language for urdu and roman urdu. IEEE 2020, Vol. 8, 91213–91226. [Google Scholar] [CrossRef]
  32. Ngo, V.M.; Cao, T.H. Semantic Search by Latent Ontological Features. New Generation Computing 2012, 30, 53–71. [Google Scholar] [CrossRef]
  33. Ngo, V.M.; Cao, T.H. A Generalized Vector Space Model for Ontology-Based Information Retrieval. Vietnamese Journal on Information Technologies and Communications 2009, 22, 43–53. [Google Scholar]
  34. Ngo, V.M.; Cao, T.H.; Le, T. WordNet-Based Information Retrieval Using Common Hypernyms and Combined Features. In Proceedings of the Proceedings of the 5th International Conference on Intelligent Computing and Information Systems (ICICIS 2011). ACM, 2011; pp. 1–6. [Google Scholar]
  35. Ngo, V.M.; Cao, T.H. Discovering Latent Concepts and Exploiting Ontological Features for Semantic Text Search. In Proceedings of the Proceedings of the 5th Int. Joint Conf. on Natural Language Processing (IJCNLP 2011). ACL, 2011; pp. 571–579. [Google Scholar]
  36. Ngo, V.M. Discovering Latent Information by Spreading Activation Algorithm for Document Retrieval. International Journal of Artificial Intelligence & Applications 2014, 5, 23–34. [Google Scholar]
  37. Dao, P.Q.; Roantree, M.; Ngo, V.M. Enhanced Dual Transformer Contrastive Network for Multimodal Sentiment Analysis. In Proceedings of the Proceedings of the 17th International Conference on Management of Digital EcoSystems (MEDES’25). Springer, 2026, CCIS, pp. 1–14.
  38. Dao, P.Q.; Ngo, V.M. Exploring, Investigating and Exploiting Sentiment Analysis Systems. Preprints.org 2025, 1–23. [Google Scholar] [CrossRef]
Figure 1. Process of methodology for hybrid model.
Figure 1. Process of methodology for hybrid model.
Preprints 203075 g001
Figure 2. A convolutional neural network.
Figure 2. A convolutional neural network.
Preprints 203075 g002
Figure 3. Illustration of a LSTM block [30].
Figure 3. Illustration of a LSTM block [30].
Preprints 203075 g003
Figure 4. Phrases related to sexual abuse on the dark web.
Figure 4. Phrases related to sexual abuse on the dark web.
Preprints 203075 g004
Figure 5. Average Precision, F1-score and AUC of baseline ML and DL models.
Figure 5. Average Precision, F1-score and AUC of baseline ML and DL models.
Preprints 203075 g005
Figure 6. Confusion Matrices for models.
Figure 6. Confusion Matrices for models.
Preprints 203075 g006
Figure 7. ROC curves for models.
Figure 7. ROC curves for models.
Preprints 203075 g007
Table 1. Average model performance across 5-fold cross-validation.
Table 1. Average model performance across 5-fold cross-validation.
Preprints 203075 i001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2026 MDPI (Basel, Switzerland) unless otherwise stated