Preprint
Article

This version is not peer-reviewed.

A Bearing Fault Diagnosis Model Based on Simplified WDCNN and RF

A peer-reviewed article of this preprint also exists.

Submitted:

25 December 2024

Posted:

26 December 2024

You are already at the latest version

Abstract
Bearings play a crucial role in the complex mechanical systems of ships, and their operational status is closely related to vibration signals. Therefore, analyzing bearing signals plays an important role in the field of fault diagnosis. In order to solve the problems of low accuracy and slow response speed in fault diagnosis through vibration signals at mixed speeds, this paper introduces an improved SWDCNN-RF model on traditional WDCNN. It was verified through the publicly available dataset of ball bearings from Western Reserve University in the United States. It was found that the improved model increased speed by 38.51% and accuracy from 97.5% to 99.6% at epoch=50, and also achieved faster convergence and smaller fluctuations during training. This study is of great significance for determining the occurrence time and type of bearing faults, and provides criteria for reliability evaluation and fault diagnosis of equipment using bearings.
Keywords: 
;  ;  ;  

1. Introduction

Maritime transportation plays a very important role in today’s economic development, and the reliability of complex mechanical systems on board ships is particularly important. With the rapid development of science and technology, fault diagnosis is an important part of reliability assessment. Traditionally, fault diagnosis of maritime machinery is carried out through physical inspection, sensor based monitoring, and the expertise of experienced maritime engineers. The recent advent of digital technologies and the emergence of big data have opened up new avenues for innovation in maritime machinery fault diagnosis [1]. Any malfunction within these bearings has the potential to cause a complete mechanical system failure, which could lead to catastrophic accidents [2]. The health and performance of rotating machinery heavily rely on the condition of the bearings that support the rotating shaft. The condition of bearings is significantly impacted by their inherent nonlinearity, causing nonlinear vibrations that affect machinery health and performance [3]. Therefore, fault diagnosis technology for rolling bearings holds significant importance in the industrial sector, and the research on bearing fault diagnosis has been booming over the past years [4]. The fault diagnosis process for ball bearings involves collecting vibration signals and extracting fault characteristics. Initially, this involves techniques such as acoustics [5], electrical current [6], acoustic emission [7], and vibration [8]. Vibration signals, being the longest-standing and most effective method, are widely employed to detect abnormalities in rolling bearings [9,10]. By attaching acceleration sensors to the bearing seat, the vibration characteristics of rolling bearings can be accurately and efficiently measured.
The development of fault diagnosis technology has evolved from traditional signal processing methods to modern machine learning and deep learning methods. Traditional methods such as time-domain analysis, frequency-domain analysis, and time-frequency analysis can initially identify the type and extent of faults by extracting and analyzing the characteristics of vibration signals. With the advancement of computing power and data processing technology, machine learning methods such as support vector machines, decision trees, and random forests, as well as deep learning methods such as convolutional neural networks (CNNs), are increasingly being used in fault diagnosis. These methods can automatically extract complex features from a large amount of vibration data, improving the accuracy and efficiency of fault diagnosis. CNN can be widely used in many fields of machine learning algorithms, and its core is image recognition. Nowadays, more and more research shows the role of CNN algorithm in fault diagnosis. So far, there are two main directions for the application of CNN network in fault diagnosis. One is to use an image sensor to capture image data of diagnostic targets on a regular basis, such as explicit images, temperature values, and so on. This method is commonly used in engineering in the direction of power systems; Another more common method is to treat CNN as a feature extractor, combined with data preprocessing and other auxiliary networks for fault diagnosis [11]. For example, VMD is used to decompose the original bearing vibration signal into a series of modal components containing fault features, and CNN is used to extract the fault features, while LSTM is used to extract the time series information of the fault signal [12].A new correlation analysis algorithm for motion steady-state signals and collected signals is proposed in reference [13], which combines CNN with FFT for motion aging. As in reference [14], the original vibration of the training bearing is first processed using the Hilbert-Huang transform to construct a new nonlinear degradation energy index. Then, CNN technology is used to identify and extract the difference between the extracted degradation energy index and the original vibration of the trained bearing. Hidden mode, used to predict the degradation life of bearings. It can be seen that CNN networks combined with preprocessed raw data or other deep learning techniques can complete fault diagnosis [15].Therefore, it can be seen that the current mainstream research basically uses vibration signals to study single faults, while there is still much room for development in using speed signals for fault diagnosis. As in reference [16], a deep learning algorithm that takes into account the motor speed is proposed in the literature. The vibration signal is regarded as an image, and then the CNN algorithm is used for fault diagnosis. In recent years, with the continuous progress of science and technology, data-driven fault diagnosis methods based on machine learning have gradually become mainstream in the field of fault diagnosis [17].
However, unlike 2D and 3D widely used in image recognition, for 1D CNNs that process one-dimensional vibration data, all stacked small convolution kernels can lead to a structure that is too deep. To address this issue, an improved deep CNN model with a wide first layer kernel (WDCNN) was proposed and successfully applied to fault diagnosis [18].The first layer uses a wide convolution kernel, while other layers use a series of small kernels, which can achieve a balance between structure and feature extraction [19].The Random Forest (RF) classifier is one of the most powerful classification algorithms [20].The main idea behind the RF algorithm is to use a combination of randomized trees and predict by majority vote among all generated decision trees [21].This in turn provides high accuracy and robustness in dealing with different problems by reducing the variance of classification and achieving lower generalization error [22].This article will combine WDCNN and random forest and simplify the structure of WDCNN to achieve the purpose of rolling bearing fault diagnosis.

2. SWDCNN-RF Basic Theory

The fault diagnosis process can be broken into three key stages: signal acquisition, feature extraction, and fault pattern recognition. Traditional approaches to fault diagnosis, while commonly used, tend to be slow, less accurate, and require a high degree of precision for data classification. These limitations make them less effective for diagnosing real-world rolling bearing faults.
The Wide Convolutional Neural Network (WDCNN) is a deep learning model designed specifically for one-dimensional data processing. It utilizes wider convolutional kernels and larger strides to efficiently capture broad, significant features from input signals. This is particularly useful for extracting meaningful patterns from one-dimensional vibration signals. By combining wide kernels and large strides, WDCNN not only enhances the model’s ability to capture local features, but it also reduces the number of convolutional layers and overall parameters. As a result, the model becomes computationally more efficient. The incorporation of batch normalization and pooling layers further accelerates the training process, improves stability, and reduces both data dimensionality and computational complexity, all of which help to mitigate the risk of overfitting.
Building on this foundation, this paper introduces an optimized model called SWDCNN-RF, which merges a simplified version of the WDCNN with Random Forest (RF) for enhanced fault diagnosis. In this approach, the original WDCNN’s five-layer convolutional structure is simplified to just two layers. To compensate for any potential loss of information due to this reduction, frequency domain features are introduced, unlike the original model, which relied solely on time-series data. This addition of frequency domain data increases the richness of the input information and helps reduce noise, improving the model’s generalization ability.
In addition, an encoder layer is incorporated to extract critical features from the input data, which may include both one-dimensional features and rotational speed. These refined features are then passed into a Random Forest algorithm. Compared to traditional convolutional neural networks, Random Forest provides superior classification performance and faster computational speed. By combining the feature extraction power of SWDCNN with the classification accuracy of Random Forest, the SWDCNN-RF model achieves a highly efficient and accurate diagnostic solution.
The proposed rolling bearing fault diagnosis model, based on the Simple Wide Convolutional Neural Network (SWDCNN), is depicted in Figure 1. This model consists of two wide convolutional layers, two max-pooling layers, a batch normalization layer, a dropout layer, two fully connected layers, an encoder layer, a Random Forest classifier, and a Softmax output layer. This architecture ensures computational efficiency while maintaining high diagnostic accuracy, making it well-suited for practical industrial applications.
The first convolutional layer employs a wide convolutional kernel (size 64) with a large stride (16), enabling it to capture a broad spectrum of features in the input vibration signal. This layer is followed by a Batch Normalization layer and a ReLU activation function, which together speed up the training process and enhance model stability. Next, a max pooling layer (with a pooling window size of 2) is applied to reduce data dimensionality and computational complexity. The second convolutional layer uses a smaller kernel size of 3, with a stride of 1, and the input data is appropriately padded. This layer also includes a batch normalization layer and a ReLU activation function, followed by another max pooling layer to further decrease data dimensionality. After these convolutional and pooling operations, the features are flattened into one-dimensional vectors, and a Dropout layer (with a dropout rate of 0.2) is applied to prevent overfitting. The model then passes through two fully connected layers—the first with 32 neurons and the second serving as an encoder, reducing feature dimensionality to 8 neurons. The output layer utilizes a Softmax activation function to convert the encoder layer’s output into a class probability distribution for fault classification.
Convolutional neural networks (CNNs), a fundamental aspect of deep learning, excel in handling both one-dimensional and multidimensional data, enabling the automatic extraction of features from raw data to complete recognition tasks. In the context of rolling bearing fault diagnosis, employing the SWDCNN-RF model eliminates the need for manual feature extraction, thereby significantly improving fault classification accuracy. This chapter provides an overview of the SWDCNN structure and the foundational principles of Random Forest (RF), before explaining how the SWDCNN-RF model is constructed to address the challenges of processing mixed data under varying operational conditions. This combined approach enhances diagnostic accuracy, reduces fault detection time, and boosts overall efficiency.
In processing one-dimensional vibration signals, the convolutional layer is a critical component of SWDCNN. By utilizing wide convolutional kernels and larger strides, this layer captures a broad range of features from the input signal, allowing for a reduction in the number of layers and parameters, which in turn improves computational efficiency. The convolutional operations focus on extracting local signal features, such as frequency components and amplitude changes, which are key to effective feature extraction in fault diagnosis tasks. This method ensures that critical signal characteristics are identified and used to improve fault recognition, making the diagnostic process faster and more accurate.
Typically following the convolutional layer, the pooling layer reduces data dimensionality and computational complexity. Max pooling is used within this layer to extract the maximum value within each pooling window, retaining the most critical features and minimizing the risk of overfitting. Additionally, the pooling layer improves the translation invariance of features, which is particularly important when processing vibration signals, as fault features may shift along the time axis. Figure 2 illustrates the max pooling process.
After the pooling operation, the size of the feature map is (M, Fs), where Fs is the number of filters in the layer and M is calculated by the following formula:
M = W F S + 1
In the formula, W represents the length and width of the input image, F represents the size of the convolution kernel, and S represents the moving step size.
The batch normalization layer is applied after the convolutional layer to standardize the input for each subsequent layer. By implementing batch normalization, SWDCNN can speed up the training process while enhancing the model’s stability and performance. This layer helps address the issue of internal covariate shift, ensuring that the input distribution for each layer remains relatively stable. As a result, the model converges more quickly and achieves higher accuracy. Batch normalization not only boosts the training efficiency but also strengthens the network’s generalization ability. The specific steps of batch normalization are as follows:
Calculate the mean μ and variance σ 2 of the current small batch;
μ = 1 m i = 1 m x i
σ 2 = 1 m i = 1 m x i μ 2
(2) Calculate the standardized   x i ' , subtract the mean μ of the small batch from the input, and then divide by the standard deviation σ 2 + ε , where ε is a very small constant to prevent the phenomenon of dividing by 0;
x i ' = x i μ σ 2 + ε
Enlarge x i ' , multiply by the same scaling value γ i , and add a bias value β i to enhance the network expression ability after batch normalization layer.
y i = γ i × x i ' + β i
After the convolution and pooling operations, the extracted signal features are flattened into one-dimensional vectors. These vectors are then fed into a fully connected layer, which further refines and combines the features using linear transformations and nonlinear activation functions like ReLU. In SWDCNN, multiple fully connected layers are typically employed, each containing a different number of neurons. This layered architecture allows the model to capture more complex, higher-level features, enhancing its ability to represent intricate patterns within the data.The fully connected layers work synergistically with the Softmax classifier, which is responsible for translating the extracted and refined features into distinct categories. By combining the feature extraction capabilities of the convolutional and pooling layers with the decision-making power of the fully connected layers and the Softmax classifier, SWDCNN achieves a high level of classification accuracy and expressive power, enabling it to handle complex tasks with greater precision.
The output of the final pooling layer is flattened into a one-dimensional feature vector, which is then fed into the fully connected layer, where it is fully connected to both the input and output. The output layer utilizes the Softmax function to convert the input neurons into a probability distribution, summing to 1, which predicts the likelihood of each bearing state. The forward propagation calculation formula for the fully connected layer is as follows:
z l + 1 j = i = 1 n W i j l α l i + b j l
Where z l + 1 is the logits value of the j-th neuron in the (l+1)-th layer, and W i j l is the weight between the i-th neuron in the first layer and the j-th neuron in the (l +1) -th layer, b j l is the bias value of all neurons in first layer to the j-th neuron in layer (l+1)-th. Figure 3 is a schematic diagram of the fully connected layer.
The expression for the ReLU function is:
f x = x , x > 0 0 , x 0 = max { 0 , x }
Draw a curve graph based on the function expression as shown in Figure 4. From Figure 3, it can be seen that when the input is less than or equal to 0, the output of the ReLU function remains constant at 0; When the input is greater than 0, the output value is equal to the input value, that is, when the input is a positive number, the derivative of the ReLU function is 1, and the gradient vanishing problem is solved.
The primary role of the encoder layer is to reduce the complexity of high-dimensional feature vectors by distilling the most relevant features for further analysis, especially when paired with Random Forest (RF) models. In SWDCNN, this layer typically utilizes fewer neurons (for instance, reducing from 32 to 8 neurons) to create a more streamlined model architecture and decrease the computational load. This reduction is achieved by applying linear transformations along with activation functions like ReLU, which map the high-dimensional features into a more manageable, lower-dimensional space. This simplification not only facilitates subsequent tasks like classification or regression but also enhances processing speed and efficiency.
Random Forests, on the other hand, are adept at optimizing their performance by automatically fine-tuning hyperparameters during training, ensuring the best possible classification results. When combined with the encoder layer, the SWDCNN-RF model excels in both computational efficiency and improved generalization. This synergy significantly elevates classification accuracy while minimizing the risk of overfitting.
As the final component of the model, the output layer plays a crucial role in delivering the ultimate predictions or classifications. In SWDCNN, the output layer typically uses a Softmax activation function to transform the encoder’s output into a probability distribution over the potential classes. For binary classification tasks, this layer consists of two neurons, each representing the likelihood of one of the two possible outcomes. In multi-class classification, the number of neurons in the output layer corresponds to the number of categories, enabling the model to produce accurate predictions for each class.

3. Fault Diagnosis of Rolling Bearings Based on SWDCNN-RF

3.1. Dataset Introduction

To verify the feasibility of the proposed model, the bearing dataset from Western Reserve University in the United States will be used for feasibility validation [23].
The experimental setup for diagnosing faults in rolling bearings, as shown in Figure 5, utilizes a test bench provided by Case Western Reserve University. This bench is equipped with accelerometers to monitor vibration signals, and the bearings tested are 6205-2RJEM SKF deep groove ball bearings. The sampling frequency of the setup is 48 kHz, adhering to the sampling theorem’s requirement that the sampling frequency must be at least twice the maximum frequency of the signal to ensure accurate data capture. This ensures that the obtained data is reliable and suitable for analysis, meeting the necessary experimental standards.
The experiment was conducted at four different rotational speeds: 1797 r/min, 1772 r/min, 1750 r/min, and 1730 r/min. To simulate various fault conditions, fault diameters of 0.007 inches, 0.014 inches, and 0.021 inches were introduced. The study evaluated four operating conditions: inner ring fault, rolling element fault, outer ring fault, and normal condition. These conditions were used for both validation and testing. The data was split into training and test sets, with a ratio of 7:3, ensuring an adequate dataset for model training and performance evaluation.
By analyzing the vibration data at these four speeds, the signal and frequency spectra of the original normal, spherical fault, inner ring fault, and outer ring fault can be obtained, as shown in Figure 6.

3.2. Application of Methods

Figure 7 shows the fault diagnosis flowchart of the model proposed in this chapter. Firstly, the training set is input into the built model for training. When the model reaches the set expected goals, the test set is input into the model for fault diagnosis, and the model outputs the diagnostic results of the test set.
Firstly, the CWRU rolling bearing dataset is preprocessed, including the labeling of fault modes, the combination of fault data at different rotational speeds, and denoising of vibration data. Next, as previously mentioned, the dataset is divided into training and testing sets at a ratio of 7:3.
For the training set, the SWDCNN model is established, and parameters are initialized. The model initialization includes the name, input shape, and number of classes as parameters. During initialization, placeholders for the loss function, optimizer, and evaluation metrics are defined. The model can take one-dimensional signals with an input shape of (1, 1024). The first convolutional layer transforms the input into 16 channels with a kernel size of 64 and a stride of 16. The padding is calculated using ‘self.clac_padding’, resulting in an output shape of (16, 128). Batch normalization is then applied to the 16 channels, with both input and output shapes being (16, 128). The first activation layer uses the ReLU function. The first pooling layer has a window size of 2, halving the length. Thus, the input shape of (16, 128) is reduced to an output shape of (16, 64). The second convolutional layer takes the output from the first pooling layer, resulting in an output shape of (32, 64), with 32 channels and a kernel size of 3, stride of 1. Padding is again calculated using ‘self.clac_padding’. Following this, the second batch normalization layer normalizes the 32 channels, maintaining the shape (32, 64). The second activation layer also employs the ReLU function. The second pooling layer, with a window size of 2, further halves the length, changing the input shape from (32, 64) to an output shape of (32, 32). The flatten layer converts the convolutional output into a one-dimensional vector, resulting in an output shape of (1, 1024). Dropout is applied, randomly discarding 20% of the neurons to prevent overfitting. There are three fully connected layers: the first fully connected layer outputs 32 neurons, the second fully connected layer (encoder layer) outputs 8 neurons, and the third fully connected layer (output layer) produces neurons corresponding to the number of classes, with shapes of (1, 2) or (1, num_classes). The Softmax layer computes the probability distribution for each class using the Softmax function. The specific input and output sizes are detailed in Table 1. After constructing the model, training is conducted to observe convergence. If convergence is not achieved, model parameters are readjusted. Upon achieving convergence, the model’s performance is evaluated to ensure it meets the fault diagnosis requirements. If the requirements are not met, further parameter adjustments are made. Once the model satisfies the requirements, the test set is input, and the SWDCNN-RF model is trained to obtain fault modes and recognition results.
The implementation code is as follows:
Model initialization section:
class sWDCNN(BasicTorchModel):
def __init__(self, name, input_shape=(1, 1024), num_classes=2):
super().__init__()
self.model_name = name
self.loss_fn = None
self.optimizer = None
self.metric_fn = None
The first convolutional layer:
padding_1 = self.clac_padding(input_shape[1], 64, 16)
self.conv1d_1 = nn.Conv1d(input_shape[0], 16, kernel_size=64, stride=16, padding=padding_1)
self.bn_1 = nn.BatchNorm1d(16)
self.active_1 = nn.ReLU()
self.pool1d_1 = nn.MaxPool1d(2)
The second convolutional layer:
input_size_2 = int(np.floor(np.ceil(input_shape[1] / 16) / 2))
padding_2 = self.clac_padding(input_size_2, 3, 1)
self.conv1d_2 = nn.Conv1d(16, 32, kernel_size=3, stride=1, padding=padding_2)
self.bn_2 = nn.BatchNorm1d(32)
self.active_2 = nn.ReLU()
self.pool1d_2 = nn.MaxPool1d(2, 2)
Flattening and fully connected layer:
self.flatten = nn.Flatten()
self.dropout = nn.Dropout(0.2)
input_size_3 = int(np.floor(input_size_2 / 2) * 32)
self.linear_3 = nn.Linear(input_size_3, 32)
self.active_3 = nn.ReLU()
Encoder and output layer:
self.linear_4 = nn.Linear(32, 8)
self.active_4 = nn.ReLU()
self.linear_5 = nn.Linear(8, num_classes)
self.active_5 = nn.Softmax(dim=1)

3.3. Experimental Comparative Analysis

This article will compare the SWDCNN-RF model and the traditional WDCNN model separately, and use both models to process the same dataset. The specific parameters of the SWDCNN-RF model are shown in Table 2, and the specific parameters of the traditional WDCNN model are shown in Table 3.
WDCNN is a relatively mature algorithm that performs well in fields such as fault diagnosis, image recognition, and speech processing. WDCNN, with its powerful feature extraction ability and good generalization performance, is widely used in the state monitoring and fault diagnosis of mechanical equipment. By processing data such as vibration signals and acoustic signals, WDCNN can accurately identify fault types, improve equipment operation reliability and maintenance efficiency. When we used the multi condition dataset as input, we still achieved a good accuracy of 97.5% in the test set, demonstrating its strong adaptability and robustness. Upon closer observation of the training process, we can observe that although high accuracy was ultimately achieved, there was a certain degree of fluctuation in the training accuracy curve and loss curve. This fluctuation may be caused by the complexity and diversity of the data. And the training epoch is 50, taking 148 seconds, which is also too long. The accuracy and running time of the model are shown in Figure 8.
We propose a new algorithm, SWDCNN-RF, which optimizes the original Wide Convolutional Neural Network (WDCNN) by simplifying its structure. The original five-layer convolutional network has been reduced to two layers, streamlining the model and improving computational efficiency. To mitigate any potential information loss from this reduction, we introduced frequency domain features, whereas the original WDCNN relied solely on one-dimensional temporal data. By adding frequency domain features, the model’s information dimension is expanded, reducing noise interference and enhancing its generalization capability.
An encoder layer has also been incorporated to extract crucial features from the input data, including one-dimensional features such as rotational speed. These refined features are then processed by the Random Forest algorithm, which has a stronger classification ability than a CNN and offers faster computational speed. The integration of the encoder and Random Forest allows for more effective classification and faster model performance.
The training process for the SWDCNN-RF algorithm was conducted over 50 epochs, with a total training time of 91 seconds, representing a 38.51% reduction in time. The model’s accuracy also improved significantly, increasing from 97.5% to 99.6%, and convergence occurred more quickly during the training process. The enhanced performance and reduced computational time are depicted in Figure 9, which illustrates both the accuracy and running time improvements achieved with the SWDCNN-RF model.
In addition, by observing the t-SNE visualization of the model outputs of the two algorithms, as shown in Figure 10, it is not difficult to see that SWDCNN-RF is clearer than WDCNN in the distance between the faults and normal conditions of the three types of ball bearings, and the confusion of the two fault modes will be less seen on SWDCNN-RF. SWDCNN-RF utilizes the ensemble learning characteristics of random forests to better capture complex features in data, increasing the robustness and generalization ability of the model. In contrast, traditional WDCNN models may have insufficient feature extraction or overfitting issues when dealing with different fault modes, resulting in unclear boundaries between different fault modes and increasing the confusion rate of classification. Through the t-SNE visualization method, we can intuitively see the distribution of SWDCNN-RF in the feature space, which can more effectively distinguish different types of fault states. This further verifies the superiority of SWDCNN-RF in handling ball bearing fault diagnosis problems. This improvement not only significantly improves accuracy, but also demonstrates stronger advantages in the interpretability and practicality of the model, especially in practical industrial applications, providing higher guarantees for the accuracy and reliability of fault diagnosis.
As illustrated in Figure 11, the confusion matrix comparing the diagnostic results of two convolutional models on the test set reveals key differences. The traditional WDCNN model shows minor errors in predicting the spherical fault mode, while the improved SWDCNN-RF model performs exceptionally well in predicting all three types of fault modes and normal operating conditions, with only minimal errors. This indicates a significant improvement in accuracy with the SWDCNN-RF model compared to the traditional WDCNN.The confusion matrix visualization highlights that the WDCNN model experiences some degree of misclassification in detecting spherical fault modes, revealing its limitations in feature extraction and differentiating between different fault types. In contrast, the SWDCNN-RF model demonstrates nearly flawless classification across various fault modes and normal conditions, showcasing its superior generalization ability and robustness when dealing with complex vibration signal data.A deeper analysis suggests that the marked improvement in classification accuracy with the SWDCNN-RF model stems from its ability to leverage the strengths of both Wide Convolutional Neural Networks (WDCNN) and Random Forests (RF). This hybrid approach enhances the model’s ability to extract and differentiate critical features, resulting in higher diagnostic precision and a more reliable performance in real-world fault diagnosis scenarios.

4. Conclusions

To enhance the speed and accuracy of ball bearing fault recognition at varying speeds, this paper presents a ball bearing fault diagnosis model based on an improved Wide Convolutional Neural Network (WDCNN). The model was validated using the ball bearing dataset provided by Western Reserve University, leading to the following key conclusions:
1) Improved Processing Speed: At 50 training epochs, the SWDCNN-RF model demonstrated a 38.51% faster response speed than the traditional WDCNN model when processing vibration signal data at mixed speeds. This significant improvement in processing speed, coupled with reduced training fluctuations, highlights the model’s ability to maintain stable and efficient performance when dealing with complex and diverse input data. This is highly valuable for real-time fault diagnosis in practical industrial applications.
2)Enhanced Diagnostic Accuracy: The SWDCNN-RF model significantly improved the diagnostic accuracy of convolutional neural networks when processing the rolling bearing dataset of Western Reserve University, with an accuracy of 99.6%. This result indicates that the model has excellent performance in handling rolling bearing fault diagnosis tasks, and can accurately identify and classify different types of faults. This high-precision diagnostic capability is of great significance for predicting and preventing equipment failures, extending equipment lifespan, and reducing maintenance costs.
3)Strong Generalization and Robustness: The superior performance of the SWDCNN-RF model is not only reflected in speed and accuracy, but also in its strong generalization ability and robustness. The model performs well under different working conditions and can effectively cope with various noise and interference factors, which makes its application prospects in industrial environments very broad. By combining the advantages of wide convolutional neural networks and random forests, the SWDCNN-RF model not only demonstrates powerful capabilities in the field of fault diagnosis, but also provides an efficient and reliable solution for other similar time series analysis tasks.

Acknowledgments

This study was supported by the National Key R&D Program of China (funding number 2022YFB2602303); Beidou +5G shipyard logistics engineering vehicle intelligent control system development (funding number KY23ZG11).

References

  1. Yu Guo, Jundong Zhang, Bin Sun, Yongkang Wang,A universal fault diagnosis framework for marine machinery based on domain adaptation,Ocean Engineering,Volume302,2024,117729,ISSN00298018. [CrossRef]
  2. Yubin Lin;Shiqing Huang;Bingyan Chen;Dawei Shi;Zewen Zhou;Rongfeng Deng;Baoshan Huang;Fengshou Gu;Andrew D. Ball.A novel drum-shaped metastructure aided weak signal enhancement method for bearing fault diagnosis[J].Mechanical Systems and Signal Processing,2024,Vol.209: 111077.
  3. Prashant H. Jain;Santosh P. Bhosle.Mathematical modeling, simulation and analysis of non-linear vibrations of a ball bearing due to radial clearance and number of balls[J].Materials Today: Proceedings,2023,Vol.72(3): 927-936.
  4. Hao Li, Jing Lin, Zongyang Liu, Jinyang Jiao, Boyao Zhang,An interpretable waveform segmentation model for bearing fault diagnosis,Advanced Engineering Informatics,Volume61,2024,102480,ISSN14740346. [CrossRef]
  5. Qin C, Wang D, Xu Z, Tang G. Improved empirical wavelet transform for compound weak bearing fault diagnosis with acoustic signals. Applied Sciences. 2020 Jan 18;10(2):682.
  6. Muhammad Irfan, Abdullah Saeed Alwadie, Adam Glowacz, Muhammad Awais, Saifur Rahman, Mohammad Kamal Asif Khan, Mohammad Jalalah, Omar Alshorman, Wahyu Caesarendra.A Novel Feature Extraction and Fault Detection Technique for the Intelligent Fault Identification of Water Pump Bearings[J].Sensors (Basel, Switzerland),2021,Vol.21(12): 4225.
  7. Xu, Guanji;Hou, Dongming;Qi, Hongyuan;Bo, Lin.High-speed train wheel set bearing fault diagnosis and prognostics: A new prognostic model based on extendable useful life(Article)[J].Mechanical Systems and Signal Processing,2021,Vol.146: 107050.
  8. Chandrabhanu Malla;Isham Panigrahi.Review of Condition Monitoring of Rolling Element Bearing Using Vibration Analysis and Other Techniques(Article)[J].Journal of Vibrational Engineering and Technologies,2019,Vol.7(4): 407-414.
  9. Jinde Zheng;Haiyang Pan;Jinyu Tong;Qingyun Liu.Generalized refined composite multiscale fuzzy entropy and multi-cluster feature selection based intelligent fault diagnosis of rolling bearing[J].ISA Transactions,2022,Vol.123: 136-151.
  10. Zhao, Zhibin;Wang, Shibin;Wong, David;Wang, Wendong;Yan, Ruqiang;Chen, Xuefeng.Fast Sparsity-Assisted Signal Decomposition With Nonconvex Enhancement for Bearing Fault Diagnosis[J].IEEE-ASME TRANSACTIONS ON MECHATRONICS,2022,Vol.27(4): 2333-2344.
  11. Dong-Jin Choi;Ji-Hoon Han;Sang-Uk Park;Sun-Ki Hong.Data Preprocessing Method in Motor Fault Diagnosis Using Unsupervised Learning[A].2019 19th International Conference on Control, Automation and Systems (ICCAS)[C],2019.
  12. Ran Song;Quan Jiang.Application of VMD Combined with CNN and LSTM in Motor Bearing Fault[A].2021 IEEE 16th Conference on Industrial Electronics and Applications (ICIEA)[C],2021.
  13. Ji-Hoon Han;Dong-Jin Choi;Sang-Uk Park;Sun-Ki Hong.A Study on Motor Poor Maintenance Detection Based on DT-CNN[A].2019 19th International Conference on Control, Automation and Systems (ICCAS)[C],2019.
  14. Yuan, Zhuang;Zhang, Laibin;Duan, Lixiang;Li, Tao.Intelligent Fault Diagnosis of Rolling Element Bearings Based on HHT and CNN[A].2018 Prognostics and System Health Management Conference (PHM-Chongqing)[C],2018.
  15. Yuan Xie;Tao Zhang.Feature extraction based on DWT and CNN for rotating machinery fault diagnosis[A].Chinese Control and Decision Conference[C],2017.
  16. Ji-Hoon Han;Dong-Jin Choi;Sang-Uk Park;Sun-Ki Hong.A Study on Fault Classification System Based on Deep Learning Algorithm Considering Speed and Load Condition[A].2019 22nd International Conference on Electrical Machines and Systems (ICEMS)[C],2019.
  17. Shan H,Jiang H,Lin, Y,; et al. A novel method for intelligent fault diagnosis of rolling bearings using ensemble deep auto encoders[J].Mechanical Systems and Signal Processing,2018(102):278-297.
  18. Zhang, W.; Peng, G.; Li, C.; Chen, Y.; Zhang, Z. A New Deep Learning Model for Fault Diagnosis with Good Anti-Noise and Domain Adaptation Ability on Raw Vibration Signals. Sensors 2017, 17, 425. [Google Scholar] [CrossRef] [PubMed]
  19. Y. Jiang et al., “Actuator Fault Diagnosis in Autonomous Underwater Vehicle Based on Principal Component Analysis,” 2019 IEEE Underwater Technology (UT), Kaohsiung, Taiwan, 2019, pp. 1-5. [CrossRef]
  20. Manuel Fernández-Delgado, Eva Cernadas, Senén Barro, and Dinani Amorim. 2014. Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 15, 1 (January 2014), 3133–3181.
  21. Lakshmanaprabu, S.K., Shankar, K., Ilayaraja, M. et al. Random forest for big data classification in the internet of things using optimal features. Int. J. Mach. Learn. & Cyber. 10, 2609–2618 (2019). [CrossRef]
  22. Z. Chai and C. Zhao, “Enhanced Random Forest With Concurrent Analysis of Static and Dynamic Nodes for Industrial Fault Classification,” in IEEE Transactions on Industrial Informatics, vol. 16, no. 1, pp. 54-66, Jan. 2020. [CrossRef]
  23. Smith W., A., Randall R. B.. Rolling element bearing diagnosis using the case western reserve university data: a benchmark study[J]. Mechanical Systems and Signal Processing, 2015, Vol. 64: 100-131.
Figure 1. Schematic diagram of model structure.
Figure 1. Schematic diagram of model structure.
Preprints 144106 g001
Figure 2. Diagram of maximum pooling operation.
Figure 2. Diagram of maximum pooling operation.
Preprints 144106 g002
Figure 3. Schematic diagram of the fully connected layer.
Figure 3. Schematic diagram of the fully connected layer.
Preprints 144106 g003
Figure 4. ReLU activation function.
Figure 4. ReLU activation function.
Preprints 144106 g004
Figure 5. Rolling bearing fault simulation test bench.
Figure 5. Rolling bearing fault simulation test bench.
Preprints 144106 g005
Figure 6. Fault Signal and Spectrum Diagram.
Figure 6. Fault Signal and Spectrum Diagram.
Preprints 144106 g006
Figure 7. Diagnostic Process.
Figure 7. Diagnostic Process.
Preprints 144106 g007
Figure 8. WDCNN Accuracy and Running Time.
Figure 8. WDCNN Accuracy and Running Time.
Preprints 144106 g008
Figure 9. Accuracy and running time of SWDCNN-RF.
Figure 9. Accuracy and running time of SWDCNN-RF.
Preprints 144106 g009
Figure 10. t-SNE visualization.
Figure 10. t-SNE visualization.
Preprints 144106 g010
Figure 11. Confusion Matrix.
Figure 11. Confusion Matrix.
Preprints 144106 g011
Table 1. .
Table 1. .
Network Layer Input size output size
Input layer (1, 1024) (1, 1024)
Convolutional layer 1 (1, 1024) (16, 128)
Batch normalization layer 1 (16, 128) (16, 128)
Activation Layer 1 (ReLU) (16, 128) (16, 128)
Pooling layer 1 (16, 128) (16, 64)
Convolutional layer 2 (16, 64) (32, 64)
Batch normalization layer 2 (32, 64) (32, 64)
Activation Layer 2 (ReLU) (32, 64) (32, 64)
Pooling layer 2 (32, 64) (32, 32)
Exhibition layer (32, 32) (1, 1024)
Dropout (1, 1024) (1, 1024)
Fully connected layer 1 (1, 1024) (1, 32)
Activation Layer 3 (ReLU) (1, 32) (1, 32)
Fully connected layer 2 (encoder) (1, 32) (1, 8)
Activation Layer 4 (ReLU) (1, 8) (1, 8)
Fully connected layer 3 (output layer) (1, 8) (1, 2)
Softmax (1, 2) (1, 2)
Table 2. .
Table 2. .
Num Network Layer Kernel Size / Stride (Neurons) Number of Filters (Kernels) Output Size (Width × Depth) Zero Padding
1 1D Convolution 1 64×1/16×1 16 128x16 Yes
2 Batch Normalization 1 16 - 128x16 No
3 1D Pooling 1 2×1/2×1 - 64x16 No
4 1D Convolution 2 3×1/1×1 - 64x32 Yes
5 Batch Normalization 2 32 32 64x32 No
6 1D Pooling 2 2×1/2×1 - 32x32 No
7 Fully Connected 1 1024 - 32x1 No
8 Fully Connected 2 (Encoder) 32 - 8x1 No
9 Fully Connected 3 8 - 4x1 No
10 Softmax 4 - 4 No
Table 3. .
Table 3. .
Num Network Layer Kernel Size / Stride (Neurons) Number of Filters (Kernels) Output Size (Width × Depth) Zero Padding
1 Convolution 1 64×1/16×1 16 128x16 Yes
2 Pooling 1 2×1/2×1 16 64 × 16 No
3 Convolution 2 3×1/1×1 32 64 × 32 Yes
4 Pooling 2 2×1 / 2 × 1 32 32 × 32 No
5 Convolution 3 3 × 1 / 1 × 1 64 32 × 64 Yes
6 Pooling 3 2 × 1 / 2 × 1 64 16 × 64 No
7 Convolution 4 3 × 1 / 1 × 1 64 16 × 64 Yes
8 Pooling 4 2 × 1 / 2 × 1 64 8 × 64 No
9 Convolution 5 3 × 1 / 1 × 1 64 6 × 64 Yes
10 Pooling 5 2 × 1 / 2 × 1 64 3 × 64 No
11 Fully Connected 100 1 100 × 1 -
12 Softmax 10 1 10 -
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated