Preprint
Article

This version is not peer-reviewed.

Hybrid Approach for Face Recognition Using Convolutional Neural Networks, Fuzzy Logic, and SVM Classifier

Submitted:

30 January 2025

Posted:

30 January 2025

You are already at the latest version

Abstract

Face recognition is a crucial area in computer vision and biometric applications, playing a significant role in security, authentication, and human-computer interaction. Despite its importance, traditional face recognition methods often encounter challenges such as variations in lighting, changes in facial expressions, and the presence of occlusions, which can affect their effectiveness. To address these limitations, hybrid approaches that combine multiple techniques have proven to enhance recognition accuracy and robustness, offering improved performance in diverse and challenging scenarios.This paper presents a hybrid approach combining Convolutional Neural Networks (CNN), Fuzzy Logic, and Support Vector Machine (SVM) classifiers for accurate and robust face recognition. The proposed model leverages the feature extraction capabilities of CNNs, the classification precision of SVMs, and the decision-making strength of Fuzzy Logic. Specifically, we use the ResNet50 CNN model to identify key facial features. The proposed method also employs Support Vector Machines (SVM) to achieve precise classification and fuzzy logic rules to refine the decision-making process by handling uncertainties and imprecise data effectively, enhancing the overall reliability of the system.Experimental results demonstrate the superiority of this hybrid model in terms of accuracy, efficiency, and robustness compared to existing techniques. Additionally, the results confirm that this hybrid approach enhances the robustness and adaptability of face recognition systems in real-world conditions. These findings highlight the potential of this integration to develop more reliable and efficient frameworks for practical applications.

Keywords: 
;  ;  ;  ;  ;  ;  ;  ;  
Subject: 
Engineering  -   Other

Introduction

Face recognition is a fundamental application in modern computer vision, widely used in areas such as security, surveillance, and access control [1] [2]. Different techniques have been proposed in this domain, and various studies have explored the relationship between Convolutional Neural Networks, fuzzy logic for face recognition [3] [4].
Traditional face recognition methods have been instrumental in the evolution of facial recognition technology, laying a strong foundation for contemporary advancements. These methods are broadly categorized into holistic, local, and hybrid approaches. Holistic methods, such as Principal Component Analysis (PCA) and Linear Discriminant Analysis (LDA), analyze the entire face as a single entity to extract global features [5]. Independent Component Analysis (ICA) further expanded these methods by considering statistical independence to enhance feature representation [6]. Local feature-based techniques, including Local Binary Patterns (LBP) and Gabor filters, focus on analyzing specific regions of the face, such as the eyes, nose, and mouth, to improve robustness against variations like partial occlusions [7] [8]. Hybrid approaches combine the strengths of holistic and local methods, aiming to balance global and localized feature analysis [9].
The advantages of these methods include their computational efficiency and effectiveness in controlled environments. Local and hybrid methods, in particular, demonstrated robustness to noise and partial occlusion, making them suitable for certain practical applications. However, these approaches have notable limitations. They are highly sensitive to variations in illumination, pose, and expression, which significantly impact performance in real-world scenarios [10]. Furthermore, traditional methods struggle to handle large, diverse datasets effectively and are vulnerable to spoofing attacks, such as the use of photographs [11].
Face recognition systems have garnered significant attention from researchers employing fuzzy logic due to its ability to handle uncertainty and imprecise data, which are prevalent in such systems. Variations in facial features—such as contours, expressions, and lighting conditions—can challenge traditional deterministic approaches. Fuzzy logic offers a flexible framework to model and interpret these variations, enabling more robust recognition in dynamic, real-world scenarios.
In particular, fuzzy logic can be applied during the classification and matching stages of face recognition. By assigning degrees of similarity rather than binary decisions, it enhances accuracy, especially when faces are partially occluded, distorted, or viewed under varying angles and lighting conditions.
In this context, Bahreini et al. [12] developed a fuzzy logic-based facial emotion recognition software using the CK+ database, supplemented with additional data from 10 participants, totaling 1,000 facial expressions. Their method involves extracting facial landmarks, calculating cosine values for facial triangles, and applying FURIA fuzzy rules. The software achieved an average accuracy of 83.2%, with recognition rates varying across emotions—the highest being for happiness (93.3%) and the lowest for fear (43.8%).
With the same objective, Yadav et al. [13] proposed the QIntTyII-DLM method, which combines quaternion representation, interval type-II fuzzy logic, and deterministic learning for face recognition. The method was evaluated on the AR, Georgia Tech, Indian Faces (female), and Faces 94 (male) datasets, showing significant error rate reductions compared to other techniques. It preserves structural information from color channels, suppresses redundancy, and effectively handles non-linearity arising from variations in appearance, occlusion, and pose. Results demonstrated a 15–20% error rate reduction for AR, 8–10% for Georgia Tech, and 3–5% for Indian Faces compared to DLM. The main drawback is its dependence on the footprint of the uncertainty parameter in the fuzzy logic step.
Furthermore, AD Algarni et al. [14] introduced two novel frameworks for generating cancellable face templates: CFR1, which combines intuitionistic fuzzy logic with random projection, and CFR2, which employs a homomorphic transform followed by random projection. Both frameworks aim to create secure and revocable biometric templates by distorting and encrypting original facial features. The approaches were evaluated using multiple datasets (ORL, FERET, LFW) and various performance metrics. Results demonstrated effective discrimination, with correlation scores exceeding 0.3 for authorized users and below 0.05 for unauthorized users. The proposed methods exhibited robustness against various attacks while maintaining strong recognition performance. However, a potential drawback is the computational complexity of the encryption processes, which may impact real-time performance in large-scale biometric systems. Additionally, the multi-stage transformation approach could increase sensitivity to noise or variations in input images, potentially affecting recognition accuracy in challenging real-world conditions.
In addition, Balovsyak et al. [15] developed a face mask recognition system combining the Viola-Jones method with fuzzy logic. The system detects facial features using Haar cascades and employs Mamdani-based fuzzy logic with triangular membership functions to distinguish between masked and unmasked faces. Implemented in Python with OpenCV and Scikit-Fuzzy libraries, it operates on the Google Colab platform and is designed for use in security and public spaces. While effective, the system is sensitive to lighting conditions and image quality, reducing accuracy in variable environments. Additionally, its reliance on the Viola-Jones method limits its performance with occlusions and unconventional mask types.
Hybrid methods, which combine techniques from holistic and feature-based approaches, have been proposed using Gabor wavelet features in combination with various subspace methods [16,17,18]. In these methods, Gabor kernels with multiple orientations and scales are convolved with an image, and the resulting outputs are concatenated into a feature vector. The feature vector is then down sampled to reduce its dimensionality. In [16], the feature vector was subsequently processed using the enhanced linear discriminant model introduced in [19]. In [17], the downsampled feature vector underwent PCA followed by ICA, and a probabilistic reasoning model from [19] was utilized to classify whether two images belonged to the same subject. In [18], kernel PCA with polynomial kernels was used to encode high-order statistics in the feature vector. These hybrid methods demonstrated improved accuracy compared to using Gabor wavelet features alone.
LBP (Local Binary Pattern) descriptors have been a significant component in many hybrid methods. In [20], an image was divided into nonoverlapping regions where LBP descriptors were extracted at multiple resolutions from each region. The LBP coefficients for each area were then concatenated into regional feature vectors. These regional feature vectors were subsequently projected onto PCA (Principal Component Analysis) combined with LDA (Linear Discriminant Analysis) subspaces. This method was further extended to color images in [21], incorporating additional color channel information to enhance feature extraction.
Laplacian PCA, an advanced version of PCA, demonstrated superior performance compared to standard PCA and kernel PCA when applied to LBP (Local Binary Pattern) descriptors in [22]. This enhanced method, Laplacian PCA, effectively captured the local variations and relationships within the data, leading to better discrimination between patterns. In [23], two novel versions of LBP were introduced: three-patch LBP (TPLBP) and four-patch LBP (FPLBP). These new LBP variants were combined with LDA (Linear Discriminant Analysis) for dimensionality reduction and SVMs (Support Vector Machines) for classification, further improving the accuracy of the hybrid approach. The proposed TPLBP (Three-Patch LBP) and FPLBP (Four-Patch LBP) descriptors can enhance face recognition accuracy by encoding similarities between neighboring patches of pixels.
More recently, [24] introduced a high-dimensional face representation by densely extracting multi-scale LBP (MLBP) descriptors around facial landmarks. In their experiments, [24] demonstrated that extracting high-dimensional features significantly boosts face recognition accuracy, increasing it by 6-7% when the dimensionality is increased from 1K to 100K dimensions. However, the main drawback of this approach is the substantial computational resources required to perform dimensionality reduction on such a large scale, which can be computationally expensive and time-consuming. Hybrid methods integrate the strengths of both holistic and feature-based approaches. However, a key limitation is selecting appropriate features that effectively capture the necessary information for face recognition. Some methods address this challenge by combining various types of features, while others incorporate a learning phase to enhance the discriminative power of these features. Deep learning techniques, which will be discussed next, advance these concepts by training end-to-end systems capable of learning a large set of features that are optimal for the recognition task.
Convolutional Neural Networks (CNNs) are the most widely used deep learning method for face recognition. The main advantage of deep learning methods like CNNs is their ability to learn a robust face representation from large amounts of data, which can handle variations such as illumination, pose, facial expressions, and age. Instead of manually designing features for these variations, CNNs learn them directly from the training data. However, the main limitation is their requirement for very large datasets to generalize effectively to new, unseen samples. Fortunately, several large-scale face datasets have recently been released publicly, providing the necessary data for training CNN models [25,26,27,28,29,30,31].
The concept of using neural networks for face recognition is not new. An early approach in 1997, the Probabilistic Decision Based Neural Network (PBDNN) [32], was proposed for tasks like face detection, eye localization, and recognition. The PBDNN used one fully connected subnet per training subject to reduce hidden units and prevent overfitting. Two PBDNNs were trained using intensity and edge features, with their outputs combined for final classification. Another early method involved combining a self-organizing map (SOM) with a convolutional neural network [33].
A self-organizing map (SOM) [34] is an unsupervised neural network that projects input data onto a lower-dimensional space while preserving the topological properties of the input space—inputs that are close together remain close in the output space. Neither the PBDNN [32] with edge features nor the SOM-based approach [33] were trained end-to-end and utilized shallow neural network architectures. A later proposal introduced an end-to-end face recognition CNN in [35].
In this context, Megahed et al. [36] proposed a hybrid CNN-fuzzy approach for adaptive e-learning, where a CNN classifies facial expressions, and a fuzzy system determines learning levels. Using two custom datasets—a 1,735-instance emotional state corpus and a 72-instance learning activity corpus—their approach demonstrated improved performance on the FER-2013 dataset and enabled adaptive learning flows based on emotional and behavioral responses. However, limitations include potential inaccuracies in facial expression recognition, reliance on facial cues while ignoring other factors, and uncertainties introduced by the fuzzy system. Privacy concerns arise from continuous video monitoring, and the system oversimplifies emotional responses, failing to account for cultural and individual differences. Additionally, the model’s accuracy (67.15%) and its assumption of a direct link between facial expressions and learning effectiveness leave room for improvement.
With the same objective, Jamal Kh-Madhloom et al. [37] developed a smile detection system integrating CNN and fuzzy logic to improve accuracy in human face analysis. The system, implemented in Python with TensorFlow and Scikit-Fuzzy, analyzes key facial features such as lips and eyes to detect smiles, with applications in biometrics, security, and forensic analysis. The CNN extracts features automatically, while fuzzy logic refines detection for ambiguous data. Despite its improved accuracy, the system's performance is affected by sensitivity to lighting and face orientation. Additionally, unspecified model size and computational demands may limit its use in resource-constrained environments. Among machine learning methods, SVMs are noted for their effectiveness in handling high-dimensional data and establishing optimal decision boundaries.
The described deep learning techniques, particularly Convolutional Neural Networks (CNNs), represent a significant advancement for face recognition tasks. Their ability to learn robust face representations from large datasets allows them to handle variations such as illumination, pose, and facial expressions without the need for manually designed features. However, their main limitation is the requirement for very large datasets to generalize well to new, unseen samples. Early methods like the Probabilistic Decision Based Neural Network (PBDNN) and self-organizing maps (SOMs) utilized shallow architectures and did not train end-to-end, which limited their effectiveness. In contrast, the introduction of an end-to-end CNN significantly improved face recognition by directly optimizing both feature extraction and classification processes, leading to higher accuracy and robustness. Despite these advances, CNNs still face challenges, such as requiring extensive computational resources and being sensitive to input quality, which can affect their performance in practical applications.
This paper introduces a novel face recognition method that combines Support Vector Machine (SVM) as the primary classifier with ResNet50 for robust feature extraction. Fuzzy logic is incorporated to refine the prediction process, improving classification accuracy. SVM effectively distinguishes known faces, while ResNet50 extracts deep, discriminative features. Fuzzy logic enhances decision-making by addressing uncertainties and variations in data, such as lighting and pose changes, thereby optimizing system performance. SVM's ability to find the optimal hyperplane for class separation improves generalization, while the integration with ResNet50 enables the use of meaningful facial representations. The addition of fuzzy logic allows better management of low-confidence predictions and challenges such as misclassification of unknown faces and overlapping feature spaces. This hybrid approach offers improved robustness and accuracy in face recognition tasks.
The rest of this report is organized as follows. Section 2 presents the proposed face recognition method. Section 3 discusses the experimental results, and Section 4 provides the conclusion.

Proposed Method

Face recognition is a biometric technique used to identify individuals based on the unique features of their faces. It provides an automated method for authentication and identification by analyzing facial patterns. These systems are reliable, fast, and effective, making them suitable for use in various environments, such as educational institutions and public spaces, where they play a critical role in enhancing security and improving organizational management [38].
In this work, we are interested in identifying individuals based on their facial features. The proposed system is conceptually different and explores new strategies. Specifically, it investigates the potential benefits of integrating advanced Convolutional Neural Networks (CNNs) for feature extraction, Support Vector Machines (SVMs) for classification, and fuzzy logic for refining prediction accuracy. By combining these powerful techniques, the system aims to improve the overall robustness and effectiveness of face recognition, particularly in challenging conditions such as varying lighting, facial expressions, or partial occlusion.
Figure 1 presents the structure of the proposed face recognition method using a hybrid approach. The key steps include image acquisition, where face images are captured for analysis; preprocessing, which normalizes and enhances the input data; feature extraction, to identify and isolate critical facial features; SVM training, to build a robust classification model; fuzzy logic integration, where fuzzy logic rules are applied to refine predictions; classification and matching, to assign identities based on the trained model; and finally, visualization and evaluation, to assess system performance and present results. Each step is crucial for enhancing the system's adaptability and robustness under varying conditions.
The proposed recognition model is structured in several stages to optimize face recognition performance. First, it employs pre-trained Convolutional Neural Networks (CNNs) for feature extraction, leveraging deep learning models to capture rich, discriminative facial features from input images. In the next stage, Support Vector Machines (SVMs) are used for classification, where they separate different classes (i.e., known individuals) based on the extracted features. Finally, fuzzy logic is incorporated to refine the decision-making process by addressing uncertainties and ambiguous data, such as variations in lighting, pose, or expression. This multi-stage approach enhances the system’s robustness and accuracy, enabling it to effectively handle challenging recognition scenarios.
Figure 2 illustrates the flowchart of the proposed face recognition system. The process begins with the input of facial data, which undergoes pre-processing to ensure it is properly formatted and cleaned for analysis. Next, the data is divided into two sets: training and testing. The training set is passed through ResNet50, a pre-trained Convolutional Neural Network (CNN), to extract relevant and deep features from the facial images. These features are then used to train a Support Vector Machine (SVM) classifier, which learns to distinguish between different individuals based on the facial characteristics.
To further enhance the system’s decision-making, fuzzy logic is applied. This step refines the output by handling uncertainties and ambiguities, particularly in situations where the data is noisy or incomplete. Once the system performs face recognition, it identifies whether the input corresponds to a known individual or an unknown person. If a known person is identified, their name is displayed as the output.

1.1.1. Input Data and preprocessing

The initial step in our face recognition system involves collecting a comprehensive dataset of facial images, which can be sourced from curated databases, live video streams, or static camera captures. In our application, publicly available datasets, such as the Labeled Faces in the Wild (LFW), are used because they offer a diverse range of facial variations, including differences in expressions, lighting conditions, and occlusions. These datasets serve as a vital foundation by providing labeled data essential for training supervised learning models and conducting thorough performance evaluations [39].
In our application, pre-processing plays a critical role in standardizing image data and reducing variability, thereby ensuring the optimal performance of the recognition system. As a first step, all input images are resized to dimensions of (224 × 224) using bi-cubic interpolation. This resizing process ensures uniformity across the dataset, making the images suitable for model input.
Subsequently, pixel values are normalized to a standardized range, typically between 0 and 1. This normalization process is crucial as it enhances computational efficiency by reducing the dynamic range of the input data, which helps prevent numerical instability during model training. By scaling the pixel values, the system ensures that the model converges more effectively and avoids issues such as exploding or vanishing gradients, ultimately contributing to more stable and accurate training outcomes.
Figure 3 displays a collection of datasets, including LFW, which are used for benchmarking and evaluating the performance of our proposed method. These datasets consist of diverse samples and are integral for training, testing, and validating the proposed computational model.
The LFW dataset consists of 13,233 face images, which are divided into training, validation, and testing sets. Specifically, 60% of the images from each class are allocated for training, 20% for validation, and 20% for testing. The dataset is utilized for a binary classification task, distinguishing Know and Unknown person. To enhance the training data, data augmentation techniques are applied. One such technique is Noise Injection, where random Gaussian noise is applied to the images. This involves adding small, random variations in pixel intensity values across the image, mimicking sensor imperfections or environmental noise during data acquisition.
In addition to these steps, alignment techniques are applied to the images to ensure that facial features, such as eyes, nose, and mouth, are consistently positioned across samples. This alignment significantly enhances the accuracy of feature extraction by reducing spatial variations caused by differences in pose, scale, or orientation. Collectively, these pre-processing techniques establish a solid foundation for accurate and reliable face recognition.
The MediaPipe Face Detector is an advanced machine-learning framework developed by Google as part of its MediaPipe library [40]. It facilitates real-time detection and localization of human faces in images and videos, making it an efficient and scalable solution for various applications. In our work, the MediaPipe Face Detector is employed on input images to accurately localize facial regions. Figure 4 presents the face detection using MediaPipe Face Detector. The preprocessing step here served to reduce the area used for the classification, i.e., consists of isolating the face area from its surroundings.

1.1.1. Features extraction using CNN

In this study, the use of Convolutional Neural Network (CNN) models such as ResNet50 for feature extraction is a standard approach in facial recognition tasks, owing to their ability to learn high-level, discriminative features from facial images. ResNet50, which is pre-trained on large datasets like ImageNet, benefits from its convolutional layers that capture complex structural patterns and spatial hierarchies within the facial images. These layers enable the model to detect key facial features such as the eyes, nose, and mouth, as well as more abstract representations of identity. The pooling layers of ResNet50 then process the extracted features, reducing their dimensionality and creating compact representations that are easier to analyze while preserving important information.
This architecture significantly improves classification performance by allowing the model to be more resilient to common variations in facial appearance, such as changes in lighting, head pose, and facial expressions. This robustness helps the system generalize effectively across diverse real-world recognition scenarios, where faces may appear in different environmental conditions or with slight distortions. Figure 5 provides a detailed illustration of the ResNet50 architecture, highlighting the various layers and their roles in the feature extraction process [41].
ResNet-50 is structured with a total of 16 processing blocks and incorporates two distinct types of shortcut connections, which are critical for its deep learning capabilities. As depicted in Figure 2(a), these shortcut connections enable the model to bypass certain layers, allowing the network to maintain gradient flow during backpropagation, which helps mitigate the issue of vanishing gradients and facilitates the learning of deeper layers. This architectural choice allows ResNet-50 to effectively learn complex features and maintain its predictive accuracy, even with the increase in depth.
One of the shortcut connection types in ResNet-50 is called the convolution block, which introduces a convolutional layer within the shortcut path. This module is used when the input dimension is smaller than the output dimension, as shown in Figure 2(b). The convolutional layer in the shortcut path adjusts the dimensions of the input to match those of the output, allowing for a smooth flow of information through the network. This helps maintain the integrity of the features as they are passed through the network, ensuring that the model can learn and adapt effectively, even as it processes deeper layers and more complex features.
Figure 5. The ResNet-50-based architecture. (a) Overview of the overall structure of the network. The symbols of ×2, ×3, ×5 in the figure is the number of blocks. (b) Structure of a convolution block where input dimension varies. (c) Structure of an identity block where input dimension does not change.
Figure 5. The ResNet-50-based architecture. (a) Overview of the overall structure of the network. The symbols of ×2, ×3, ×5 in the figure is the number of blocks. (b) Structure of a convolution block where input dimension varies. (c) Structure of an identity block where input dimension does not change.
Preprints 147762 g005
The other type of shortcut connection in ResNet-50 is called the identity block, where the input and output dimensions are the same. In this module, there is no convolutional layer in the shortcut path, as shown in Figure 2(c). The identity block simply allows the input to be passed directly to the output, preserving the original dimensions. This type of shortcut connection helps maintain the flow of information through the network without altering its representation, making it effective for situations where the input and output features do not require transformation. This design enables the network to learn residual mappings more efficiently and contributes to the overall success of deep networks like ResNet-50.
Each module in ResNet-50 features a bottleneck building block structure, which consists of three convolution layers: 1 × 1, 3 × 3, and 1 × 1. This design helps reduce the number of parameters in the network, thereby improving computational efficiency, while still maintaining high performance. In this study, we retrained all layers of the network using a CXR (Chest X-ray) image dataset, which included four categories for classification. To achieve this, we fine-tuned the network without freezing any layers, allowing the entire network to adapt to the specific characteristics of the dataset. Additionally, the last fully-connected layer and the final classification layer were replaced to match the number of categories in the classification task, ensuring the network could output the correct number of class labels.
Since ResNet requires input data to be sized at 224 × 224, all input images were resized using bi-cubic interpolation [42]. We performed 10-fold cross-validation for network re-training. Out of the total 1,200 images, 1,080 images (90% of the total) were used for re-training, while the remaining 120 images were used for validation.
The average of these classification accuracies is taken as the overall accuracy of the network. The mini-batch size was set to 36, and the optimizer used was Adam. Adam is an optimization algorithm that serves as an alternative to the classical stochastic gradient descent method, iteratively updating network weights based on training data [43]. During retraining, to improve accuracy, the parameters were adjusted to increase the learning rate in the newly replaced fully connected layer. Conversely, the learning rate was reduced in the transfer layer. Additionally, the learning rate was decreased every 5 epochs. To prevent overfitting, L2 norm regularization was applied to the cost function (also referred to as the loss function). The epoch setting was determined by evaluating accuracy at each iteration cycle, and retraining was stopped after 5 consecutive iterations without improvement in accuracy.

1.1.1. Classification using SVM

Support Vector Machine (SVM) is a supervised learning algorithm widely employed for binary classification tasks [44]. It works by identifying an optimal hyperplane that maximally separates data points from different classes, ensuring a clear decision boundary with the largest possible margin.
In this study, feature maps extracted from the ResNet50 convolutional neural network were flattened into one-dimensional vectors to serve as input to the SVM classifier. A grid search technique, combined with cross-validation, was applied to identify the optimal SVM hyperparameters.
The fundamental equation for a Support Vector Machine (SVM) in binary classification is the decision function that determines the class of a given input. It is expressed as:
f ( x ) = s i g n ( w T ϕ ( x ) + b )
where: w is the weight vector, which defines the orientation of the hyperplane. ϕ ( x ) is the transformation function that maps input features x into a higher-dimensional space (in cases where the kernel trick is used). b is the bias term, which shifts the hyperplane. f ( x ) is the output, which determines the class of the input xxx based on the sign (+1 or -1).
For training, the objective is to minimize the following optimization problem:
m i n w , b , ξ 1 2 w 2 + C i = 1 n ξ i
subject to:
y i w T ϕ x i + b 1 ξ i ,           ξ i 0 i = 1 , 2 , , n
where:
y i represents the true label of the i t h training sample (+1 or −1). ξ i is the slack variable, allowing some misclassification for non-linearly separable data. C represents the regularization parameter, controlling the trade-off between maximizing the margin and minimizing the classification error.
By using kernel functions such as the Radial Basis Function (RBF) kernel, the algorithm efficiently computes the separation in higher-dimensional spaces without explicitly transforming the data.
This process involved selecting the kernel type (Radial Basis Function, or RBF), the regularization parameter (C), and the kernel coefficient (γ), which are critical for controlling the model's complexity and the shape of the decision boundary.
The RBF kernel is defined as:
K x i x j = e x p ( γ x i x j 2 )
where x i and x   j are feature vectors, and γ > 0 determines the influence of a single training example. The regularization parameter C controls the trade-off between achieving a low error on the training data and maintaining a large margin. These parameters are tuned to optimize the model's performance.
The SVM was trained using these feature vectors derived from the training dataset. Once trained, the classifier's performance was evaluated on the test dataset, providing insights into its accuracy and generalization capability. This rigorous parameter tuning and evaluation ensured that the SVM model achieved optimal performance for the given task.

1.1.1. Fuzzy rules

Support Vector Machines (SVMs) classify data by identifying optimal hyperplanes in feature spaces. While highly effective, SVMs may struggle with borderline cases or data exhibiting significant overlap between classes. Integrating a fuzzy inference system provides a mechanism to handle such uncertainties by introducing graded classifications and refining outputs.
Preprints 147762 i001Preprints 147762 i002
Membership functions are utilized to map input variables into fuzzy sets effectively. In this implementation, Prediction Confidence, which reflects the confidence level of SVM predictions and ranges from low to high, is categorized into three fuzzy sets: Low, Medium, and High. Similarly, Class Overlap, which quantifies the degree of overlap between classes for a given input, is divided into three fuzzy sets: Low, Moderate, and High. To ensure smooth transitions between these categories, triangular and trapezoidal functions are commonly employed to represent the fuzzy sets. The four main components of the fuzzy inference system (FIS) are depicted in the algorithm 1.
By combining SVM outputs with uncertainty metrics, the fuzzy inference system ensures that decisions are informed by contextual information. This integration reduces misclassifications and enhances the model’s adaptability to complex datasets. Integrating a fuzzy inference system with SVM effectively handles uncertainties and imprecise data. This approach ensures smooth transitions across decision boundaries, robust classification of borderline cases, and improved prediction accuracy, thereby enhancing the reliability of the overall system.

Experimental Results

The study implemented the proposed method for Face recognition, as illustrated in Figure 2. This method utilizes feature vectors derived from training images as inputs, producing classification information for each data point as output.
In this section, we first introduce the dataset and evaluation metrics employed in the study. The performance of the proposed method is then demonstrated through a series of experiments, including quantitative comparisons, misclassification analysis, parameter analysis, computational complexity evaluation, and generalization tests.

1.1. Dataset

To evaluate the proposed method, LFW datasets were utilized in this study. The Labeled Faces in the Wild (LFW) dataset is a benchmark dataset widely used for the task of face verification. Introduced in 2007, it was designed to challenge face recognition systems with real-world images that include variations in lighting, pose, expression, and occlusion. The dataset contains a collection of face images from the internet, making it diverse in terms of age, gender, and ethnicity. The LFW dataset is particularly important for evaluating face verification algorithms, which determine whether two given images belong to the same person or not.
LFW datasets contains 13,233 images of 5,749 individuals, taken in real-world, uncontrolled environments with variations in pose, lighting, and facial expressions. For evaluation purposes, the dataset is split into two subsets, with the second subset serving as a standard benchmark. This benchmark consists of 10 groups, each containing 300 positive and 300 negative image pairs, used for cross-validation. While the LFW dataset has made substantial contributions to the advancement of face verification techniques, it does have some limitations. One notable issue is the uneven distribution of images, with 4,096 individuals represented by just one image each. Despite these drawbacks, the LFW dataset continues to be a widely adopted benchmark, with numerous deep learning models achieving near-perfect accuracy in face verification tasks [39]. Table 1 presents the descriptions of the LFW Dataset.

1.1. Performance Measures

The proposed model was rigorously evaluated to ensure its effectiveness and reliability. Standard public face image datasets were utilized to assess the model's performance. Specifically, each dataset was divided into two parts: three folds were used for training, while the remaining fold was used for testing in each iteration. Comparative analyses with existing state-of-the-art models were also conducted to highlight the improvements introduced by the proposed method.
Performance metrics such as accuracy, precision, recall, and computational efficiency were analyzed to provide a comprehensive assessment [45][46]. Accuracy represents the proportion of correct predictions to the total number of samples. Precision is defined as the ratio of true positive samples to the total number of positive predictions. Recall is the ratio of true positive samples to the total number of samples in the positive class. The F-score, which is the harmonic mean of precision and recall, was also calculated. The formulas for these metrics are provided below:
A c c u r a c y = T P + T N T P + T N + F P + F N
Accuracy reflects the overall correctness of the model by calculating the ratio of correctly predicted instances (both true positives and true negatives) to the total instances examined.
P r e c i s i o n = T P T P + F P
Precision measures the proportion of true positive predictions relative to the total positive predictions made by the model, indicating the model's effectiveness in minimizing false positives.
F 1   s c o r e = 2 . T P 2 . T P + F P + F N
F1 score is the harmonic mean of precision and recall, providing a balanced measure that accounts for both false positives and false negatives, making it particularly useful in situations where class distribution is imbalanced.
R e c a l l = T P T P + F N
Recall, also known as sensitivity, measures the model’s ability to identify all relevant positive instances, calculated as the proportion of true positives to the total actual positives.
S p e c i f i c i t y = T N T N + F P
Specificity assesses the model's ability to correctly identify negative instances by calculating the proportion of true negatives to the total actual negatives.
Where, TP (True Positive) refers to the number of positive samples correctly predicted. FP (False Positive) represents the number of negative samples incorrectly classified as positive. TN (True Negative) is the number of negative samples correctly identified. FN (False Negative) denotes the number of positive samples mistakenly classified as negative.
By analyzing these metrics, we can gain a comprehensive understanding of the proposed model's performance and its effectiveness in accurately categorizing the data.
Quantitative comparison and analysis
The performance of the face recognition system was evaluated using the Labeled Faces in the Wild (LFW) dataset. On the LFW dataset, the competing methods consist of five widely adopted approaches, which include Bahreini et al. [47]; Megahed et al. [48], Khan et al. [49], Abdullah et al. [50] and Huang et al. [51]. A detailed comparison of their performance is provided in Table 4.
As shown in Table 2, the accuracy, precision, recall, and F-score achieved by our proposed method on the LFW dataset exceed 96%. These results highlight the robustness and effectiveness of the proposed model, demonstrating its consistent high performance in all key evaluation metrics.
The system demonstrated strong performance, achieving an overall accuracy of 92.51%, which indicates that it correctly classified over 92% of the test samples. Additionally, the weighted average precision reached 96.14%, showcasing the model's capability to deliver accurate predictions with minimal false positives. The weighted average recall was 93.2%, reflecting its effectiveness in identifying true positives across the majority of labels. Furthermore, the weighted average F1-Score also stood at 93.8%, emphasizing a well-maintained balance between precision and recall across the dataset.
Table 2 provides a comparative analysis of face recognition techniques utilized in various research studies, emphasizing the diversity in approaches and datasets. Bahreini et al. [47] implemented a fuzzy logic-based system using the FURIA algorithm on the CK+ database, achieving an accuracy of 83.2%. Similarly, Megahed et al. [48] proposed a hybrid approach combining Convolutional Neural Networks (CNN) with fuzzy logic and tested it on the FER-2013 dataset, yielding an accuracy of 67.15%. On the other hand, Khan et al. [49] adopted a machine learning approach leveraging the Principal Component Analysis (PCA) algorithm and achieved a variable accuracy 77.5% using the NCR-IT dataset. Abdullah et al. [50] applied the PCA algorithm in a real-time video streaming context, reporting an accuracy of 80%. Huang et al. [51] introduced an innovative method based on Point-to-Set Correlation Learning (PSCL), which was evaluated on the COX face database, resulting in accuracy 52.11%. These studies collectively highlight the importance of the choice of algorithm and dataset in determining the effectiveness of a face recognition model.
In contrast, the proposed method presented in this study stands out by utilizing a hybrid approach that integrates CNN, Support Vector Machines (SVM), and fuzzy logic. This methodology was tested on the Labeled Faces in the Wild (LFW) dataset, a standard benchmark in face recognition research. The proposed model achieved an impressive accuracy of 92.51%, significantly outperforming the other techniques listed in Table 2. The superior performance can be attributed to the effective combination of CNN's feature extraction capabilities, SVM's robust classification potential, and fuzzy logic's ability to handle uncertainty. This result demonstrates the advantage of hybrid models in achieving high recognition accuracy, particularly when benchmarked against state-of-the-art approaches.
In addition to quantitative evaluation, the system's predictions were analyzed through visual outputs. Figure 6 shows examples of the system’s performance.
The images above demonstrate the performance of the face recognition system. The first image represents correct predictions, where the system successfully identified the individual, showcasing the model's ability to handle well-represented and distinct classes effectively. However, the second image highlight misclassification. In this image, the actual label was incorrectly predicted as "Unknown," likely due to insufficient representation of this class in the training data and potential variations in facial features or conditions. These misclassifications emphasize the need for a more balanced and diverse dataset, as well as potential improvements in feature extraction and classification methods, to enhance the system’s performance across all classes.
The proposed face recognition system exhibited strong performance, achieving an overall accuracy of 92.51% on the LFW dataset. By combining Convolutional Neural Networks (CNNs), Support Vector Machines (SVMs), and fuzzy logic, the system effectively tackled challenges associated with real-world conditions, including variations in lighting, facial expressions, and pose. This hybrid approach demonstrates its ability to handle uncertainties and improve classification accuracy.
The system achieved outstanding results for well- represented classes, with near-perfect precision and recall for individuals with ample training data. However, classes with limited representation showed weaker performance, underscoring the importance of a more balanced dataset to enhance recognition across all categories.
Hence, the hybrid system provides a robust framework for practical applications, particularly in security, authentication, and adaptive systems. The use of fuzzy logic enhanced decision refinement, making the system more adaptable to uncertainties. However, challenges remain, including computational complexity, which could hinder real-time deployment in resource- constrained environments, and sensitivity to data imbalance, which impacts the system's generalizability to less-represented classes.

4. Conclusions

In this study, we proposed a hybrid face recognition framework integrating Convolutional Neural Networks (CNNs), Support Vector Machines (SVMs), and fuzzy logic to address challenges posed by variations in lighting, pose, and facial expressions. The system demonstrated robust performance on the Labeled Faces in the Wild (LFW) dataset, achieving a notable accuracy of 92.51%, along with high precision, recall, and F1 scores. These results highlight the effectiveness of the proposed methodology in enhancing face recognition accuracy by leveraging advanced feature extraction, classification, and decision refinement techniques. Despite its strengths, the system faced challenges with underrepresented classes, emphasizing the need for more diverse datasets.
Future work aims to address these limitations by incorporating larger, more balanced datasets to improve system adaptability. Enhancements in handling partial occlusions and integrating real-time optimizations are also critical to extend the framework’s applicability to practical scenarios. Additionally, we will improve the system’s ability to recognize unknown individuals by storing features of unrecognized faces and reintroducing them into the training process, enabling the model to recognize these individuals in future interactions. Furthermore, adopting privacy-preserving mechanisms, such as secure biometric templates, and leveraging advancements in edge AI and quantum computing could significantly enhance performance. These improvements will ensure the framework’s reliability and scalability for real-world security and authentication applications.

Abbreviations

CNNs Convolutional Neural Networks
SVMs Support Vector Machines
PCA Principal Component Analysis
LDA Linear Discriminant Analysis
ICA Independent Component Analysis
CK+ Cohn-Kanade Plus database
FURIA Fuzzy Unordered Rule Induction Algorithm
QIntTyII-DLM Quaternion Interval Type II-Based Deterministic Learning Machine
ORL Olivetti Research Laboratory
FERET Face Recognition Technology
LFW Labeled Faces in the Wild
FER Facial Expression Recognition
ResNet-50 Residual Network 50
RBF Radial Basis Function
FER-2013 Facial Expression Recognition 2013 Dataset
PCA Principal Component Analysis
PSCL Point To Set Correlation Learning

References

  1. Jha, Medha, et al. "Face Recognition: Recent Advancements and Research Challenges." 2022 13th International Conference on Computing Communication and Networking Technologies (ICCCNT). IEEE, 2022. [CrossRef]
  2. LI, Huixiang, et al. AI Face Recognition and Processing Technology Based on GPU Computing. Journal of Theory and Practice of Engineering Science, 2024, 4.05: 9-16. [CrossRef]
  3. Trigueros, Daniel Sáez, Li Meng, and Margaret Hartnett. "Face recognition: From traditional to deep learning methods." arXiv preprint arXiv: 1811.00116 (2018). [CrossRef]
  4. Ali, W., Tian, W., Din, S. U., Iradukunda, D., & Khan, A. (2021). Classical and modern face recognition approaches: a complete review. Multimedia tools and applications, 80, 4825-4880. [CrossRef]
  5. Dabhade, S. B., et al. "Face recognition using principle component analysis and linear discriminant analysis: comparative study." 2nd national conference on advancements in the era of multi-disciplinary systems AEMDS-2013. 2013.
  6. Hyvärinen, Aapo. "Independent component analysis: recent advances." Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 371.1984 (2013): 20110534. [CrossRef]
  7. Ahonen, T., Hadid, A., & Pietikäinen, M. (2006). Face description with local binary patterns: Application to face recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence, 28(12), 2037-2041. [CrossRef]
  8. Sun, Yuechuan, and Jun Yu. "Facial expression recognition by fusing Gabor and local binary pattern features." MultiMedia Modeling: 23rd International Conference, MMM 2017, Reykjavik, Iceland, January 4-6, 2017, Proceedings, Part II 23. Springer International Publishing, 2017.
  9. Ali, Waqar, et al. "Classical and modern face recognition approaches: a complete review." Multimedia tools and applications 80 (2021): 4825-4880. [CrossRef]
  10. Zhao, W., Chellappa, R., Phillips, P. J., & Rosenfeld, A. (2003). Face recognition: A literature survey. ACM Computing Surveys (CSUR), 35(4), 399-458. [CrossRef]
  11. Hadid, A., Pietikäinen, M., & Ahonen, T. (2007). A discriminative feature space for detecting and recognizing faces. IEEE Transactions on Pattern Analysis and Machine Intelligence, 29(6), 1041-1046. [CrossRef]
  12. Bahreini, Kiavash, Wim Van der Vegt, and Wim Westera. "A fuzzy logic approach to reliable real-time recognition of facial emotions." Multimedia Tools and Applications 78 (2019): 18943-18966. [CrossRef]
  13. Yadav, Sudesh, and Virendra P. Vishwakarma. "Robust face recognition using quaternion interval type II fuzzy logic–based feature extraction on colour images." Medical & Biological Engineering & Computing 62.5 (2024): 1503-1518. [CrossRef]
  14. Algarni, Abeer D., et al. "Efficient implementation of homomorphic and fuzzy transforms in random-projection encryption frameworks for cancellable face recognition." Electronics 9.6 (2020): 1046. [CrossRef]
  15. Balovsyak, Serhiy, et al. "Face Mask Recognition by the Viola-Jones Method Using Fuzzy Logic." International Journal of Image, Graphics and Signal Processing (IJIGSP) 16.3 (2024): 39-51. [CrossRef]
  16. Liu, Chengjun, and Harry Wechsler. "Gabor feature based classification using the enhanced fisher linear discriminant model for face recognition." IEEE Transactions on Image processing 11.4 (2002): 467-476. [CrossRef]
  17. Liu, Chengjun, and Harry Wechsler. "Independent component analysis of Gabor features for face recognition." IEEE transactions on Neural Networks 14.4 (2003): 919-928. [CrossRef]
  18. Liu, Chengjun. "Gabor-based kernel PCA with fractional power polynomial models for face recognition." IEEE transactions on pattern analysis and machine intelligence 26.5 (2004): 572-581. [CrossRef]
  19. Liu, Chengjun, and Harry Wechsler. "Robust coding schemes for indexing and retrieval from large face databases." IEEE Transactions on image processing 9.1 (2000): 132-137. [CrossRef]
  20. Chan, Chi-Ho, Josef Kittler, and Kieron Messer. "Multi-scale local binary pattern histograms for face recognition." Advances in Biometrics: International Conference, ICB 2007, Seoul, Korea, August 27-29, 2007. Proceedings. Springer Berlin Heidelberg, 2007.
  21. Chan, Chi-Ho, Josef Kittler, and Kieron Messer. "Multispectral local binary pattern histogram for component-based color face verification." 2007 First IEEE International Conference on Biometrics: Theory, Applications, and Systems. IEEE, 2007. [CrossRef]
  22. Zhao, Deli, Zhouchen Lin, and Xiaoou Tang. "Laplacian PCA and its applications." 2007 IEEE 11th International Conference on Computer Vision. IEEE, 2007. [CrossRef]
  23. Wolf, Lior, Tal Hassner, and Yaniv Taigman. "Descriptor based methods in the wild." Workshop on faces in'real-life'images: Detection, alignment, and recognition. 2008.
  24. Yoo, Cheol-Hwan, et al. "High-dimensional feature extraction using bit-plane decomposition of local binary patterns for robust face recognition." Journal of Visual Communication and Image Representation 45 (2017): 11-19. [CrossRef]
  25. Y. Sun, X. Wang, and X. Tang, “Deep learning face representation from predicting 10,000 classes,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1891–1898, 2014.
  26. D. Yi, Z. Lei, S. Liao, and S. Z. Li, “Learning face representation from scratch,” arXiv preprint arXiv:1411.7923, 2014. [CrossRef]
  27. O. M. Parkhi, A. Vedaldi, A. Zisserman, et al., “Deep face recognition.,” in BMVC, vol. 1, p. 6, 2015.
  28. Y. Guo, L. Zhang, Y. Hu, X. He, and J. Gao, “Ms-celeb-1m: A dataset and benchmark for large-scale face recognition,” in European Conference on Computer Vision, pp. 87–102, Springer, 2016.
  29. A. Nech and I. Kemelmacher-Shlizerman, “Level playing field for million scale face recognition,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3406–3415, IEEE, 2017.
  30. A. Bansal, A. Nanduri, C. D. Castillo, R. Ranjan, and R. Chellappa, “Umdfaces: An annotated face dataset for training deep networks,” in Biometrics (IJCB), 2017 IEEE International Joint Conference on, pp. 464–473, IEEE, 2017. [CrossRef]
  31. Q. Cao, L. Shen, W. Xie, O. M. Parkhi, and A. Zisserman, “Vggface2: A dataset for recognising faces across pose and age,” arXiv preprint arXiv:1710.08092, 2017. [CrossRef]
  32. [103] S.-H. Lin, S.-Y. Kung, and L.-J. Lin, “Face recognition/detection by probabilistic decision-based neural network,” IEEE transactions on neural networks, vol. 8, no. 1, pp. 114–132, 1997. [CrossRef]
  33. S. Lawrence, C. L. Giles, A. C. Tsoi, and A. D. Back, “Face recognition: A convolutional neural-network approach,” IEEE transactions on neural networks, vol. 8, no. 1, pp. 98–113, 1997. [CrossRef]
  34. T. Kohonen, “The self-organizing map,” Neurocomputing, vol. 21, no. 1-3, pp. 1–6, 1998.
  35. S. Chopra, R. Hadsell, and Y. LeCun, “Learning a similarity metric discriminatively, with application to face verification,” in Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE Computer Society Conference on, vol. 1, pp. 539–546, IEEE, 2005. [CrossRef]
  36. Megahed, M., C Mohammed, A. (2020). Modeling adaptive E-learning environment using facial expressions and fuzzy logic. Expert Systems with Applications, 157, 113460. [CrossRef]
  37. Madhloom, J. K., Diwan, S. A., & Abdulhussein, Z. A. (2020). Smile detection using convolutional neural network and fuzzy logic. Journal of Information Science and Engineering, 36(2), 269-278.
  38. 6. Balovsyak, S., Derevyanchuk, O., Kovalchuk, V., Kravchenko, H., & Kozhokar, M. (2024). Face Mask Recognition by the Viola-Jones Method Using Fuzzy Logic. International Journal of Image, Graphics and Signal Processing, 16(3), 39-51. [CrossRef]
  39. Huang, G. B., Mattar, M., Berg, T., & Learned-Miller, E. (2008). Labeled faces in the wild: A database forstudying face recognition in unconstrained environments. In Workshop on faces in'Real-Life'Images: detection, alignment, and recognition.
  40. Jakhete, Sumitra A., and Nilima Kulkarni. "A Comprehensive Survey and Evaluation of MediaPipe Face Mesh for Human Emotion Recognition." 2024 8th International Conference on Computing, Communication, Control and Automation (ICCUBEA). IEEE, 2024. [CrossRef]
  41. MANDAL, Bishwas; OKEUKWU, Adaeze; THEIS, Yihong. Masked face recognition using resnet-50. arXiv preprint arXiv:2104.08997, 2021. [CrossRef]
  42. Abdullah, Dahlan, et al. "Application of interpolation image by using bi-cubic algorithm." Journal of Physics: Conference Series. Vol. 1114. IOP Publishing, 2018. [CrossRef]
  43. Brownlee, J. (2021) Gentle Introduction to the Adam Optimization Algorithm for Deep Learning. Machine Learning Mastery.
  44. Sen, Pratap Chandra, Mahimarnab Hajra, and Mitadru Ghosh. "Supervised classification algorithms in machine learning: A survey and review." Emerging Technology in Modelling and Graphics: Proceedings of IEM Graph 2018. Springer Singapore, 2020. [CrossRef]
  45. Alqudah, A.; Qazan, S.; Alquran, H.; Qasmieh, I.; Alqudah, A. COVID-19 Detection from X-ray Images Using Different Artificial Intelligence Hybrid Models. Jordan J. Electr. Eng. 2020, 6, 168. [Google Scholar] [CrossRef]
  46. Alquran, H.; Alsalatie, M.; Mustafa, W.A.; Al Abdi, R.; Ismail, A.R. Cervical Net: A Novel Cervical Cancer Classification Using Feature Fusion. Bioengineering 2022, 9, 578. [Google Scholar] [CrossRef] [PubMed]
  47. Bahreini, K., Van der Vegt, W., C Westera, W. (2019). A fuzzy logic approach to reliable real-time recognition of facial emotions. Multimedia Tools and Applications, 78, 18943-18966. [CrossRef]
  48. Megahed, M., C Mohammed, A. (2020). Modeling adaptive E-learning environment using facial expressions and fuzzy logic. Expert Systems with Applications, 157, 113460. [CrossRef]
  49. Khan, A., et al.: Forensic video analysis: passive tracking system for automated Person of Interest (POI) localization. IEEE Access 6, 43392–43403 (2018). [CrossRef]
  50. Abdullah, N.A., et al.: Face recognition for criminal identification: an implementation of principal component analysis for face recognition. In: The 2nd International Conference on Applied Science and Technology 2017 (ICAST’17) (2017). [CrossRef]
  51. Huang, Z., et al.: A benchmark and comparative study of video-based face recognition on COX face database. IEEE Trans. Image Process. 24(12), 5967–5981 (2015). [CrossRef]
Figure 1. The structure of the proposed face recognition method.
Figure 1. The structure of the proposed face recognition method.
Preprints 147762 g001
Figure 2. Flowchart of the proposed Face Recognition System.
Figure 2. Flowchart of the proposed Face Recognition System.
Preprints 147762 g002
Figure 3. Examples of images from the LFW database.
Figure 3. Examples of images from the LFW database.
Preprints 147762 g003
Figure 4. Face Detection using MediaPipe: (a) Original Image, (b) Detection of Face, (c) Isolation of the interest object (Face).
Figure 4. Face Detection using MediaPipe: (a) Original Image, (b) Detection of Face, (c) Isolation of the interest object (Face).
Preprints 147762 g004
Figure 6. System performance.
Figure 6. System performance.
Preprints 147762 g006
Table 1. Descriptions of the LFW Dataset:.
Table 1. Descriptions of the LFW Dataset:.
Total Number of Images 13,233
Total Number of Individuals 5,749
Images per Individual Varies (some individuals are represented by only one image)
Image Variations Varying poses, lighting, and expressions
Evaluation Method 10 subsets for cross-validation, each containing 300 positive and 300 negative image pairs
Standard Benchmark View 2, used as the main benchmark for model evaluation
Image Source Collected from the internet, making it diverse in terms of age, gender, and ethnicity
Usage Primarily used for face verification tasks in machine learning models
Table 2. Comparison results of different face recognition methods.
Table 2. Comparison results of different face recognition methods.
Author Approach Algorithm Dataset Accuracy Precision Recall F-score
Bahreini et al. [47] Fuzzy Logic FURIA CK+ Database 83.2% 86.46% 83.82% 84.36%
Megahed et al. [48] Hybrid Approach CNN
Fuzzy Logic
FER-2013 67.15% 69.78% 67.65% 68.08%
Khan et al. [49] Machine Learning PCA algorithm NCR-IT 77.5% 80.54% 78.07% 78.58%
Abdullah et al. [50] PCA algorithm Real-time video stream 80% 83.13% 80.59% 81.16%
Huang et al. [51] Point-to-set correlation
learning (PSCL)
COX face database 52.11% 54.15% 52.49% 52.84%
Our proposed method Hybrid Approach CNN
SVM
Fuzzy Logic
LFW 92.51% 96.14% 93.18% 93.78%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated