Preprint Article Version 1 Preserved in Portico This version is not peer-reviewed

The Impact of Irrationals on the Range of Arctan Activation Function for Deep Learning Models

Version 1 : Received: 5 April 2023 / Approved: 6 April 2023 / Online: 6 April 2023 (08:55:22 CEST)

How to cite: Tümer Sivri, T.; Pervan Akman, N.; Berkol, A. The Impact of Irrationals on the Range of Arctan Activation Function for Deep Learning Models. Preprints 2023, 2023040079. https://doi.org/10.20944/preprints202304.0079.v1 Tümer Sivri, T.; Pervan Akman, N.; Berkol, A. The Impact of Irrationals on the Range of Arctan Activation Function for Deep Learning Models. Preprints 2023, 2023040079. https://doi.org/10.20944/preprints202304.0079.v1

Abstract

Deep learning has been applied in many areas that have had a significant impact on applications that improves real-life challenges. The success of deep learning in a wide range of areas is due in part to the use of activation functions, which are particularly effective at solving non-linear problems. Activation functions are a key focus for researchers in artificial intelligence who aim to improve the performance of neural networks. This article provides a comprehensive explanation and comparison of different activation functions, with a focus on the arc tangent and its variations specifically. The paper presents experimental results that show that variations of the arc tangent using irrational numbers such as pi, the golden ratio, and Euler’s number, as well as a self-arctan function, produce promising results. Since we experimented with promising activation functions on two different problems, and datasets, we reached a result that different irrationals work well for different problems. In other words, arctan ϕ gives the best results mostly for multiclass classification and arctan e gives the best results for time series prediction problems. The paper focuses on a multi-class classification problem applied to the Reuters Newswire dataset and a time-series prediction problem on Türkiye energy trade value to show the impacts of activation functions.

Keywords

Deep neural networks; Activation functions; Multi-class classification; Time-Series prediction; Reuters data; Energy trade value

Subject

Computer Science and Mathematics, Artificial Intelligence and Machine Learning

Comments (0)

We encourage comments and feedback from a broad range of readers. See criteria for comments and our Diversity statement.

Leave a public comment
Send a private comment to the author(s)
* All users must log in before leaving a comment
Views 0
Downloads 0
Comments 0
Metrics 0


×
Alerts
Notify me about updates to this article or when a peer-reviewed version is published.
We use cookies on our website to ensure you get the best experience.
Read more about our cookies here.