Submitted:
23 June 2025
Posted:
25 June 2025
You are already at the latest version
Abstract
Keywords:
Chapter 1: Introduction
1.1. Background
1.2. Problem Statement
1.3. Research Objectives
- To develop a framework that integrates contextual emotion modeling within transformer architectures, thereby improving the emotional depth and thematic coherence of generated poetry.
- To curate a comprehensive dataset of poetic texts that encompasses a range of emotional themes, poetic forms, and cultural contexts, facilitating the training and evaluation of emotion-aware models.
- To evaluate the performance of enhanced models against traditional poetry generation approaches using both quantitative metrics and qualitative assessments from poetry experts and enthusiasts.
- To contribute to the discourse on AI and creative expression, exploring the implications of emotion-driven models in the context of artistic authenticity and cultural representation.
1.4. Research Questions
- How can contextual emotion modeling be effectively integrated into transformer architectures to enhance the quality of generated poetry?
- What are the unique emotional and thematic characteristics that should be represented in a curated dataset for poetry generation?
- How do enhanced models compare to traditional approaches in terms of emotional resonance, thematic coherence, and overall quality of generated poetry?
- What ethical considerations arise from the use of AI in creative expression, particularly concerning emotional authenticity and cultural sensitivity?
1.5. Significance of the Study
1.6. Structure of the Thesis
- Chapter 1: Introduction, which outlines the background, problem statement, objectives, research questions, significance, and structure of the study.
- Chapter 2: Literature Review, providing a comprehensive overview of existing research on transformer models, poetry generation, and the integration of emotion in AI applications.
- Chapter 3: Methodology, detailing the research design, data collection methods, model development, evaluation metrics, and analytical techniques employed in the study.
- Chapter 4: Results and Discussion, presenting the findings of the model evaluations and dataset analysis, along with a critical discussion of the implications for poetry generation.
- Chapter 5: Conclusion and Future Work, summarizing key findings, discussing limitations, and proposing potential avenues for further research.
Chapter 2: Literature Review
2.1. Introduction
2.2. Poetry Generation in Natural Language Processing
2.2.1. The Evolution of Text Generation
2.2.2. The Transformer Architecture
2.2.3. Current Approaches to Poetry Generation
2.3. Emotion in Text Generation
2.3.1. The Role of Emotion in Poetry
2.3.2. Contextual Emotion Modeling
2.3.3. Integrating Emotion into Transformer Models
2.4. Challenges in Emotion-Aware Poetry Generation
2.4.1. Data Scarcity and Quality
2.4.2. Understanding Emotional Complexity
2.4.3. Balancing Creativity and Coherence
2.5. Future Directions in Emotion-Aware Poetry Generation
2.5.1. Developing Comprehensive Datasets
2.5.2. Enhancing Model Architectures
2.5.3. Evaluating Emotion-Aware Poetry
2.6. Conclusion
Chapter 3: Methodology
3.1. Introduction
3.2. Research Design
3.2.1. Dataset Development
3.2.1.1. Selection of Poetic Forms
3.2.1.2. Data Sources
- Literary Anthologies: Classic and contemporary poetry collections were utilized to gather a rich variety of poems.
- Online Repositories: Platforms dedicated to poetry, such as Poetry Foundation and Academy of American Poets, were accessed to curate modern poems.
- Community Contributions: Engagement with local poets and literary communities facilitated the collection of original works, ensuring cultural representation and authenticity.
3.2.1.3. Annotation of Emotional Dimensions
- Expert Review: Poets and literary scholars reviewed the poems to assign appropriate emotional labels based on thematic content and linguistic cues.
- Natural Language Processing Tools: Sentiment analysis tools were employed to assist in identifying emotional tone, although human oversight was crucial to ensure accuracy.
3.2.2. Model Architecture
3.2.2.1. Attention Mechanism
- Emotion-Aware Attention Layers: Modifying the attention mechanism to incorporate emotional context by adjusting the attention weights based on the emotional annotations of the input text.
- Contextual Embeddings: Integrating embeddings that represent emotional states, allowing the model to generate poetry that aligns with the intended emotional tone.
3.2.2.2. Model Training
- Transfer Learning: The pre-trained GPT model is adapted to the specific task of poetry generation by training it on the newly created dataset.
- Hyperparameter Optimization: Key hyperparameters, such as learning rate, batch size, and number of epochs, are tuned to maximize performance.
3.2.3. Evaluation of Generated Outputs
3.2.3.1. Quantitative Metrics
- BLEU Score: Measures the overlap between generated poetry and reference texts, providing insights into linguistic accuracy.
- Perplexity: Assesses the model's predictive capability, with lower values indicating better performance.
3.2.3.2. Qualitative Evaluation
- Expert Reviews: A panel of poets and literary scholars evaluated the generated poetry based on criteria such as thematic coherence, emotional resonance, and stylistic diversity. Reviews were conducted blind to minimize bias.
- User Surveys: Poetry enthusiasts participated in surveys to provide feedback on their perceptions of the generated outputs, assessing emotional engagement and aesthetic qualities.
3.3. Data Analysis
3.3.1. Statistical Analysis
3.3.2. Thematic Coding
3.4. Ethical Considerations
- Cultural Sensitivity: Engaging with diverse cultural stakeholders ensured that the dataset reflects authentic voices and experiences.
- Respect for Intellectual Property: Proper attribution was maintained for all sourced works, and community contributors were acknowledged in the dataset.
3.5. Conclusion
Chapter 4: Results and Discussion
4.1. Introduction
4.2. Model Evaluation
4.2.1. Overview of Experimental Setup
4.2.2. Dataset Characteristics
4.2.3. Evaluation Metrics
-
Quantitative Metrics:
- ∘
- BLEU Score: Measures the overlap between generated poetry and reference texts, indicating the model's ability to replicate linguistic patterns.
- ∘
- Perplexity: Assesses the model's predictive capability, with lower values indicating better performance.
- ∘
- Emotion Consistency Score (ECS): A novel metric developed for this study to evaluate the emotional alignment of generated poetry with the intended emotional context.
-
Qualitative Metrics:
- ∘
- Expert Reviews: A panel of poets and literary scholars evaluated the generated poetry based on criteria such as thematic coherence, emotional resonance, and stylistic diversity.
- ∘
- User Surveys: Feedback from poetry enthusiasts provided insights into the perceived emotional impact and overall quality of the generated outputs.
4.2.4. Results Summary
4.2.4.1. Quantitative Findings
| Model | BLEU Score | Perplexity | ECS (Emotion Consistency Score) |
| Baseline Model | 0.42 | 32.5 | 0.60 |
| Emotion-Aware Model 1 | 0.58 | 28.3 | 0.75 |
| Emotion-Aware Model 2 | 0.55 | 29.0 | 0.72 |
| Emotion-Aware Model 3 | 0.60 | 27.8 | 0.78 |
4.2.4.2. Observations
- BLEU Scores: The emotion-aware models consistently outperformed the baseline model in BLEU scores, indicating a higher degree of linguistic coherence and relevance to the reference texts. The best-performing model, Emotion-Aware Model 3, achieved a BLEU score of 0.60, a notable improvement over the baseline score of 0.42.
- Perplexity: The emotion-aware models exhibited lower perplexity scores, suggesting improved predictive capabilities in generating contextually appropriate poetry. Emotion-Aware Model 3 had the lowest perplexity at 27.8, demonstrating its effectiveness in understanding poetic structures.
- Emotion Consistency Score: The ECS results indicated that the emotion-aware models maintained a stronger alignment with the intended emotional themes. Emotion-Aware Model 3 achieved an ECS of 0.78, reflecting its proficiency in generating poetry that resonates emotionally with readers.
4.2.5. Qualitative Findings
4.2.5.1. Expert Reviews
4.2.5.2. User Studies
4.3. Discussion
4.3.1. Implications for Poetry Generation
4.3.2. The Role of Emotion in Creative Expression
4.3.3. Ethical Considerations
4.3.4. Future Research Directions
- Expanding Emotional Dimensions: Investigating additional emotional dimensions and their interplay in poetry generation, allowing for more nuanced outputs.
- Cross-Cultural Applications: Applying emotion-aware modeling techniques to diverse linguistic and cultural contexts to examine their effectiveness across different poetic traditions.
- Improving Evaluation Metrics: Further refining evaluation metrics to better capture the artistic and emotional qualities of generated poetry, fostering a more comprehensive understanding of model performance.
4.4. Conclusion
Chapter 5: Conclusion and Future Work
5.1. Introduction
5.2. Summary of Key Findings
5.2.1. Enhancements in Poetry Generation
5.2.2. Dataset Development
5.3. Implications for Natural Language Processing and Creative Expression
5.4. Ethical Considerations
5.5. Future Research Directions
5.5.1. Expanding the Emotion Model
5.5.2. Cross-Linguistic Studies
5.5.3. User Interaction and Feedback Loops
5.5.4. Interdisciplinary Approaches
5.6. Conclusion
References
- Shabarirajan, K. J., Logeshwar, B. S., Aadhithyan, D., & Elakkiya, R. (2024, July). Comparative Performance Analysis of Neural Architectures for Poem Generation. In 2024 International Conference on Signal Processing, Computation, Electronics, Power and Telecommunication (IConSCEPT) (pp. 1-6). IEEE.
- De la Rosa, J., Pérez, Á., De Sisto, M., Hernández, L., Díaz, A., Ros, S., & González-Blanco, E. (2023). Transformers analyzing poetry: multilingual metrical pattern prediction with transfomer-based language models. Neural Computing and Applications, 1-6. [CrossRef]
- Dunđer, I., Seljan, S., & Pavlovski, M. (2020, September). Automatic machine translation of poetry and a low-resource language pair. In 2020 43rd International Convention on Information, Communication and Electronic Technology (MIPRO) (pp. 1034-1039). IEEE.
- Aepli, N. (2024). There Is Plenty of Room at the Bottom: Challenges & Opportunities in Low-Resource Non-Standardized Language Varieties (Doctoral dissertation, University of Zurich).
- Pranida, S. Z., Genadi, R. A., & Koto, F. (2025). Synthetic Data Generation for Culturally Nuanced Commonsense Reasoning in Low-Resource Languages. arXiv preprint arXiv:2502.12932.
- Meyer, J. B. (2019). Generating Free Verse Poetry with Transformer Networks (Doctoral dissertation, Reed College).
- Abdibayev, A. (2023). Probing and Enhancing the Reliance of Transformer Models on Poetic Information (Doctoral dissertation, Dartmouth College).
- Audichya, M. K., & Saini, J. R. (2023, October). ChatGPT for creative writing and natural language generation in poetry and prose. In 2023 International Conference on Advanced Computing Technologies and Applications (ICACTA) (pp. 1-7). IEEE.
- Joe IR, P., Sudheer Kumar, E., K, K., & S, S. (2025). Sentiment-aware visual verses: limerick generation from images using transformer models for therapeutic and educational support. Journal of Poetry Therapy, 1-25.
- Sheverack, R. (2021). Modern-Day Shakespeare: Training Set Experiments with a Generative Pre-Trained Transformer-Best Paper.
- Khanmohammadi, R., Mirshafiee, M. S., Rezaee Jouryabi, Y., & Mirroshandel, S. A. (2023). Prose2Poem: the blessing of transformers in translating prose to Persian poetry. ACM Transactions on Asian and Low-Resource Language Information Processing, 22(6), 1-18. [CrossRef]
- Zaki, M. Z. (2024). Revolutionising Translation Technology: A Comparative Study of Variant Transformer Models–BERT, GPT and T5. Computer Science and Engineering–An International Journal, 14(3), 15-27.
- Dakhore, M., Eti, M., Diwakar, M., Sivanantham, A., Verma, L., & Shyam, M. (2024, December). Blending the Powers of BERT and Neural Style Transfer for Artistic Text Generation in Poetry. In 2024 IEEE 2nd International Conference on Innovations in High Speed Communication and Signal Processing (IHCSP) (pp. 1-6). IEEE.
- Oghaz, M. M., Saheer, L. B., Dhame, K., & Singaram, G. (2025). Detection and classification of ChatGPT-generated content using deep transformer models. Frontiers in Artificial Intelligence, 8, 1458707. [CrossRef]
- Riaz, A., Abdulkader, O., Ikram, M. J., & Jan, S. (2025). Exploring topic modelling: a comparative analysis of traditional and transformer-based approaches with emphasis on coherence and diversity. International Journal of Electrical and Computer Engineering (IJECE), 15(2), 1933-1948. [CrossRef]
- Liu, R. (2025). The impact of generative pre-trained transformers on creative writing instruction: Enhancing student engagement and expressive competence. Journal of Computational Methods in Sciences and Engineering, 14727978251337961. [CrossRef]
- Das, A., & Verma, R. M. (2020). Can machines tell stories? A comparative study of deep neural language models and metrics. IEEE Access, 8, 181258-181292. [CrossRef]
- Thapa, D., Joe IR, P., & Anand, S. Im-to-Lim: A Transformer-Based Framework for Limerick Generation Associated with an Image. Shajina, Im-to-Lim: A Transformer-Based Framework for Limerick Generation Associated with an Image.
- Alpdemir, Y., & Alpdemir, M. N. (2024, April). AI-Assisted Text Composition for Automated Content Authoring Using Transformer-Based Language Models. In 2024 IEEE International Conference on Advanced Systems and Emergent Technologies (IC_ASET) (pp. 1-6). IEEE.
- Koziev, I., & Fenogenova, A. (2025, May). Generation of Russian Poetry of Different Genres and Styles Using Neural Networks with Character-Level Tokenization. In Proceedings of the 9th Joint SIGHUM Workshop on Computational Linguistics for Cultural Heritage, Social Sciences, Humanities and Literature (LaTeCH-CLfL 2025) (pp. 47-63).
- Novikova, S., Sagar, S., Lin, P., Li, M., & Markovic, P. English and Chinese poetry generation Software project: Deep Learning for the Processing and Interpretation of Literary Texts.
- Elzohbi, M. (2025). AlGeoRhythm: Exploring the Geometric Patterns in Poetry Rhythms and the Generation of Beat-Aligned Poetic Texts.
- Rahman, M. H., Kazi, M., Hossan, K. M. R., & Hassain, D. (2023). The Poetry of Programming: Utilizing Natural Language Processing for Creative Expression. International Journal of Advanced Research, 8, 2456-4184.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).