Preprint
Article

Global and Local Knowledge Distillation Method for Few-Shot Classification of Electrical Equipment

This version is not peer-reviewed.

Submitted:

02 May 2023

Posted:

03 May 2023

Read the latest preprint version here

A peer-reviewed article of this preprint also exists.

Abstract
With the increasing use of intelligent mobile devices for online inspection of electrical equipment in smart grids, the limited computing power and storage of these devices pose challenges for carrying large algorithm models and it’s hard to obtain a large number of images of electrical equipment in public. In this paper, we propose a novel distillation method that compresses the knowledge of teacher networks into a small few-shot classification network using a global and local knowledge distillation strategy. Central to our method is exploiting the global and local relationship between the features exacted by the backbone of the teacher network and student network. We compare our method with recent state-of-the-art methods in three public datasets and achieve the best performance. We also contribute a new dataset, EEI-100, specifically designed for classification of electrical equipment, and demonstrate that our method achieves a prediction accuracy of 94.12% with only 5-shot images.
Keywords: 
;  ;  ;  

1. Introduction

As an essential component of the power system, daily inspection of electrical equipment is imperative to ensure the secure and stable operation of the power system [1]. The conventional manual screening and analysis approach can no longer meet the escalating demand for image analysis of electrical equipment. With the advent of the smart grid, an increasing number of unmanned aerial vehicles (UAVs) are being deployed for online inspection. The application of artificial intelligence techniques for condition monitoring of electrical equipment can significantly enhance the efficiency of detection and maintenance. Image classification is a crucial prerequisite for equipment condition monitoring based on image information. For instance, to monitor the normalcy of electrical equipment such as transformers or insulators, their images must be initially distinguished.
With the advancement of deep learning in image recognition applications, various classification and recognition methods for electrical equipment images based on deep convolutional neural network (CNN) have been proposed. However, deep CNN training often relies on large-scale labeled data, which is challenging to obtain for all categories of electrical equipment due to their safety and sensitivity. Therefore, this paper adopts few-shot learning (FSL) method to the task of electrical equipment image classification. This method randomly divides the power image dataset into a base class set and a new class set, trains the model backbone on the base class set, and finally combines the model backbone with the classifier to complete the recognition training of the new class with a small number of samples.
Furthermore, power inspection heavily relies on intelligent mobile devices, such as inspection robots and UAVs. However, due to the limited storage capacity of these devices, the classification model's capacity must not be excessively large. Otherwise, it cannot be deployed on such mobile devices. To address these challenges, this paper presents a novel few-shot electrical image classification algorithm based on knowledge distillation.
Knowledge distillation [2,3] is an efficient model compression method that compresses the knowledge of teacher networks into very small student networks. Early knowledge distillation methods that minimize the KL divergence of predicted class probability distributions between student and teacher networks rely on the output of the last layer of the model and learn a limited amount of information. Recent work has begun to study the features of the middle layer of the distillation network, focusing on the learning of local features of the image. However, there are both differences in the global appearance and local details between electrical equipment, and the algorithm model must fully mine this information in the learning process in order to comprehensively represent the electrical equipment images and achieve higher classification precision. Therefore, this paper proposes a global and local knowledge distillation method for few-shot classification of electrical equipment.
In this paper, we propose three main contributions:
  • We present a novel distillation approach that compresses the knowledge of teacher networks into a compact student network, enabling efficient few-shot classification. The incorporation of global and local relationship strategies during the distillation process effectively directs the student network towards achieving performance levels akin to those of the teacher network.
  • We contribute a new dataset that contains 100 classes of electric equipment with 4000 images. The dataset contains a wide range of various electrical equipment, including power generation equipment, distribution equipment, industrial electrical equipment, and household electrical equipment.
  • We demonstrate the effectiveness of our proposed method by validating it on three public datasets and comparing it with the SOTA methods on the electric image dataset we introduced. Our proposed method outperforms all other methods and achieves the best performance.

2. Related Work

2.1. Electrical Images Classification

Image-based equipment condition monitoring has been proven to be effective in enhancing the working life of equipment and providing early failure warning. In recent years, machine learning has made significant progress in the field of image classification for electrical equipment. Bogdann presented a machine learning method for determining the state of each switch by analyzing images of the switches in power distribution substations [4].Zhang implemented FINet based on improved YOLOv5 to inspect the insulators and their defects for ensuring the safety and stability of power system [5].To address few fault cases and deficient monitoring information in transformer diagnostic tasks, Xu provides an improved few-shot learning method based on approximation space and belief functions [6]. Yi proposed a label distribution CNNs classifier to estimate the aging time of the conductor morphology of high-voltage transmission line [7].
It is noteworthy that the majority of the aforementioned investigations have concentrated on a restricted range of electrical apparatus. These models necessitate a substantial quantity of training data to guarantee optimal performance. Nevertheless, acquiring adequate electrical equipment images in a practical setting may prove to be challenging, and the proportion of labeled samples is minimal. To a certain extent, the classification of electrical equipment images does not truly belong to a big data problem. Rather, it pertains to FSL domains.

2.2. Few-shot Classification

In recent years, FSL has attracted researcher’s widespread attention in the field of computer vision and machine learning, and a large number of few-shot image classification algorithms have been proposed. Depending on the learning paradigm used, these methods can be broadly divided into two categories: meta-learning-based methods and transfer-based learning methods.
Meta-learning is a promising approach that leverages episodic training to simulate the real test environment by randomly selecting several subtasks. This enables the acquisition of meta-knowledge that facilitates the rapid identification of new categories. Based on the type of meta-knowledge learned, meta-learning methods can be classified into optimization-based and metric-based methods. Optimization-based meta-learning methods employ a two-tier optimization process to learn the optimizer for quickly processing new tasks. A well-known example of such methods is Model-Agnostic Meta-Learning (MAML) [8]. MAML obtains the optimal initialization parameters of the model through meta-training, enabling the model to adapt to new tasks after a few gradient updates. In addition, the learning rate and gradient direction are also important factors for the optimizer [9,10]. However, these methods require storage and computation of higher-order derivatives, resulting in high memory and computational costs. On the other hand, metric-based methods use nonparametric classifiers as the basic learner, avoiding the aforementioned issues. The key factors of these methods are feature extraction and similarity measurement, which offer ample room for improvement. PARN [11] proposed a feature extractor which is learning an offset for each cell in the convolution kernel to extract more efficient features, building by deformable convolutional layers. CC+rot [12] improved the transfer ability of feature extractors by adopting auxiliary self-supervised tasks. Zhang et al. [13] used the pre-trained visual saliency detection model to segment the foreground and background of the image, and then extract the foreground and background features respectively. With the proven effectiveness of attention mechanisms in extracting discriminating features, several few-shot classification (FSC) tasks have adopted this method, including CAN [14], AWGIM [15], and CTM [16]. Additionally, in metric-based meta-learning methods, the measurement of similarity is also crucial. SEN [17] combines Euclidean distance and norm distance to improve the effectiveness of Euclidean distance measurement in high-dimensional spaces. FPN [18] calculated the reconstruction error between the support sample and the query sample as the similarity score. DN4 [19] and Deep EMD [20] obtain rich similarity measures directly on local features.
Recent studies have indicated that FSC of transfer-learning method can attain comparable performance to that of meta-learning method with complex episodic training. Such methods typically combine pre-trained feature extractors on all base class datasets with arbitrary traditional classifiers to make classification decisions for query samples of unknown classes. Reference [21] showed that pre-training the entire base class dataset using the cross-entropy loss function, followed by fine-tuning the pre-trained model using support samples of the visible class, can provide a powerful baseline for FSC tasks. Since then, several works have been proposed to improve the representation performance of feature extractors. For example, Neg-Cosin [22] proposed to use the non-negative interval Cosine loss function to optimize the model, thereby increasing the distance between the training sample and its corresponding parametric prototype, which can effectively improve the generalization performance of the model. S2M2[23] used manifold mixing as an effective regularization method to improve the generalization performance of the model. Reference [24] and [25] used rotation prediction and mirror prediction as self-supervised tasks to add to the pre-training process, and experimental results show that self-supervised tasks are effective methods to improve feature representation performance.
In conclusion, many recent works have emphasized the importance of feature representation, both meta-learning-based and transfer-learning-based methods tended to employ highly complex networks to enhance feature representation. Therefore, deploying these methods to real-world applications usually occupies high computing resources (storage space, computing power, etc.) and introduces high time delays, which cannot meet the actual needs of the classification tasks of electrical equipment images. Hence, in this study, we employ the knowledge distillation-based model compression algorithm to accomplish the task of few-shot image classification to reduce the model parameters.

2.3. Knowledge Distillation

Knowledge distillation is one of the most effective model compression methods, which has garnered significant research interest in both industry and academia due to its simple training strategy and effective performance. It leverages the knowledge acquired by a teacher network with a large scale to guide the training of a small-scale student network, enabling the latter to achieve comparable performance despite having fewer parameters.
Two key elements in current knowledge distillation methods can be summarized as: (1) the definition of effective knowledge types, (2) Effective transfer of knowledge from teacher networks to student networks. The classical knowledge distillation method minimizes the KL divergence of the predicted class probability distribution between the student and teacher networks. In order to make better use of the knowledge information contained in the teacher network, the follow-up work focuses more on how to better mine the feature knowledge hidden in the middle layer of the network. For example, AT [26] proposed to take the spatial attention of the hidden layer features of the teacher network as knowledge, and instructs the student network to imitate its attention feature map. Recently, the relationship between samples features has been proposed as a more effective knowledge. RKD [27] proposed a relational knowledge distillation method, which used distance and angle to measure the relationship between samples features, as a valid type of knowledge during distillation. Peng et al. [28] used the kernel function to obtain higher-order relationships between samples features as effective distillating knowledge.
Although the model compression methods based on feature relation knowledge mentioned above can effectively improve the performance of small-capacity student networks, the current work only focuses on the local relationship between individual sample features, ignoring the global relationship between samples features. Therefore, this paper proposes a method based on global and local knowledge distillation and applies it to the task of FSC of electrical equipment images.

3. Methodology

3.1. Problem Definition

In few-shot image classification tasks, given a certain size of image dataset I , it is randomly divided into three subsets represented as I t r a i n , I val and I test respectively. I t r a i n , also known as the base dataset, is used for pre-training the classification model. Assuming that the training set has C t categories, the mth image sample is represented as x m , and its corresponding label is represented as y m . I val is used for validation, while I test , also known as the new class dataset, is used for testing. For I val , and I test , many N-way-K-shot subtasks are randomly sampled, and each task consisting of a support sample set I S and a query sample set I Q . I S is constructed by randomly selecting N categories from I val or I test , and then randomly selecting K samples from each category. The set of the nth category is denoted as I n = { ( I k , y k ) } k = 1 K , and the kth image in the nth category is denoted as I k . I Q is composed of Q samples randomly selected from each residual sample category, denoted as I Q = { I q } q = 1 Q , where I q represents the qth query sample. Therefore, the problem of few-shot image classification can be described as using the model trained on the base class dataset and the support sample set to make classification decisions for query samples

3.2. FSC Network based on Global and Local Knowledge Distillation

We propose a novel few-shot electrical image classification algorithm based on knowledge distillation. Figure 1 shows the overall architecture of our network. We first trained a high-performance teacher network through self-supervised learning, and then guided the training of the student network by the teacher network. To fully utilize the prior knowledge of the teacher network, we designed a knowledge distillation method based on global and local relationships. This method can transfer the global and local features of the images extracted by the teacher network to the student network, enabling the small-capacity student network to learn more effective features about the images and achieve better image classification.

3.2.1. Pre-train of Teacher Network

The teacher model consists of a backbone convolutional neural network and two linear classifiers. The backbone network f θ ( ) is used for feature extraction of images, one classifier L w ( · ) is used for predicting the base class of image samples, and the other classifier L r ( · ) is used for predicting the rotation category in self-supervised tasks. Additionally, each classifier is followed by a Softmax layer. M image samples are randomly selected from the base class dataset, and each image is rotated at 0°, 90°, 180°, and 270°, with the mth image represented as x m , its corresponding base class label as y m , and its corresponding rotation label as y ^ m =[0,1,2,3].
The image is fed into the teacher network, and the d-dimensional feature representation f θ ( x m ) is extracted by the backbone network. The classification scores of the base class prediction classifier and the rotation prediction classifier for the features are expressed as:
{ S m b = L w ( f θ ( x m ) ) S m r = L r ( f θ ( x m ) )
Furthermore, the aforementioned classification scores are transformed into base class and rotation class prediction probabilities through a Softmax layer, as shown follows.
{ p 1 ( y m = c | x m ) = e S m b c c = 1 C b e S m b c p 1 ( y ^ m = r | x m ) = e S m r r r = 1 4 e S m r r
where, S m b c represents the cth component of the score vector for the base class, S m r r represents the rth component of the score vector for the predicted class, C b represents the number of base class labels, and p ( y m = c | x m ) and p ( y ^ m = r | x m ) represent the probability output values of the base classifier and the rotation classifier, respectively. The cross-entropy loss function and the self-supervised loss function are calculated to obtain the training loss function, as shown follows.
L ( θ , w , r ) = m = 1 M c = 1 C b y m c log p ( y m = c | x m ) m = 1 M r = 1 4 y ^ m r log p ( y ^ m = r | x m )
Based on the loss function in Equation (3), the parameters of the teacher network are optimized to complete the pre-training process of the teacher network.

3.2.2. Global and Local Knowledge Distillation

Firstly, a student network is constructed, which consists of a backbone neural network B ϕ ( ) composed of a small number of convolutional layers and a linear classifier C H ( ) . Next, a batch of M images randomly selected from the base dataset I t r a i n are inputted into both the teacher network and the student network. The mth image is represented by feature maps z m t = f θ ( I m ) and z m s = B ϕ ( I m ) obtained from the backbone of the teacher network and the student network, respectively. Finally, the features are fed into the linear classifier to obtain the output value of the student network, as shown follows.
S m = C H ( z m t )
Furthermore, the above output classification scores are transformed into classification prediction probabilities through the Softmax layer, as shown follows.
p t ( y m = c | x m ) = e S m c c = 1 C b e S m c
The formula for calculating the cross-entropy loss function between the output values of a student network and the true labels is shown as follows.
l 1 ( ϕ , H ) = m = 1 M c = 1 C b y m c log p t ( y m = c | x m )
In order to enable students to learn the representation of global features of images by the teacher network through online learning, we adopt the maximum mean discrepancy between the feature spaces of the two networks as the global loss function, which is calculated as follows.
l 2 ( ϕ , H ) = 1 M 2 m = 1 M m ' = 1 M z m t z m ' t T + 1 M 2 m = 1 M m ' = 1 M z m s z m ' s T 1 M 2 m = 1 M m ' = 1 M z m t z m ' s T
In addition, we calculate the Euclidean distance between each sample feature in the two networks as the local loss function, and its calculation formula is as follows.
l 3 ( ϕ , H ) = 1 M m = 1 M z m t z m s 2
In summary, the total loss function for the student network is shown in equation (9). Based on equation (9), the student network is trained and the parameters in the network are updated until optimal, thereby completing the knowledge distillation process from the teacher network to the student network.
L ( ϕ , H ) = l 1 ( ϕ , H ) + α 1 l 2 ( ϕ , H ) + α 2 l 3 ( ϕ , H )

3.2.3. Few-shot Evaluation

After completing the knowledge distillation task in Section 3.2.2, the base classifiers in the student network are first removed. Then, the parameters of the backbone neural network B φ ( ) are fixed, and features are extracted from both the support and query samples. Finally, based on N-way-K-shot method, the query samples are classified using Equation (10), where the features of the kth support sample and the qth query sample are denoted as B φ ( I k ) and B φ ( I q ) , respectively, and g φ { } represents a classifier with parameters φ . Any traditional classifier can be used to complete the classification prediction task.
y ^ q = g φ { B ϕ ( I q ) | k = 1 N K B ϕ ( I k ) }

4. Experiments

4.1. Experiments on Public Datasets

4.1.1. Experiment Setup

We evaluate our knowledge distillation method on three public datasets, namely MiniImageNet, CIFAR-FS, and CUB. The experiments are conducted on a workstation equipped with NVIDIA 3090Ti GPU and implemented using Pytorch software. To ensure a fair comparison with current small sample image classification methods, a commonly used 4-layer convolutional neural network and ResNet12 are adopted as the student network and teacher network, respectively. During the training phase, we use the SGD optimizer to optimize our models in all experiments, where the momentum is set to 0.9 and weight decay is set to 5×10-4. We train for 100 epochs, with an initial learning rate of 0.025, which is reduced by half after 60 epochs. In the testing phase, we conduct 5-way-1-shot and 5-way-5-shot tests. Specifically, we randomly perform 2000 classification subtasks in the testing dataset. In each subtask, 15 images are randomly selected from each class as query images for testing. The evaluation criterion for the algorithm's classification performance is the average accuracy of all subtasks, and the standard deviation of the accuracy under a 95% confidence interval should also be provided.

4.1.2. Parametric Analysis Experiment

It can be seen from formula (9) that α 1 and α 2 are important hyperparameters in the process of distilling the student network. Firstly, the parameter α 2 is fixed at 1, and the value of parameter α 1 is varied with a step size of 0.1 within the range of [0,1]. The test accuracy of the student network under different values of α 1 is shown in Figure 2 (a) and (b). It can be seen from the results that the model performance is optimal when the value of α 1 is 0.5. Therefore, the value of α 1 is set to 0.5, and then the α 2 is varied with a step size of 0.01 within the range of [0,0.1]. The test accuracy of the student network under different values of α 2 is shown in Figure 2(c) and (d). It can be seen from the results that the optimal value for α 2 is 0.1.

4.1.3. Ablation Experiment

The innovation of this work lies in proposing a knowledge distillation algorithm for global and local relationships. In order to verify the effectiveness of the proposed method, detailed ablation experiments are conducted on three public datasets. The knowledge distillation algorithms using only global and local relationships are denoted as Global and Local, respectively, and their fusion is denoted as Global-Local. The classification accuracies of these methods on 5-way 1-shot and 5-way 5-shot tasks are shown in Table 1. The results in the table indicate that for both 5-way 1-shot and 5-way 5-shot tasks on all datasets, the classification accuracy of Global-Local is consistently higher than that of Global and Local. The experiments demonstrate that global and local relationships are complementary, and their fusion can extract richer image features. Therefore, the knowledge distillation algorithm based on global and local relationships can further improve the performance of knowledge distillation.

4.1.4. Comparison Experiment with Existing Methods

This paper compares our method with the state-of-the-art (SOTA) methods in recent years, which are mainly divided into two categories: meta-learning-based methods and transfer learning-based methods. The comparison results with these methods are shown in Table 2.
According to the results in Table 2, it can be observed that: (1) On the MiniImageNet dataset, our proposed method achieves the best classification performance. Compared with the best performing method in the meta-learning-based category, HGNN, our method outperforms it by 2.23% and 0.9% on 1-shot and 5-shot classification tasks, respectively. In the transfer learning-based category, compared with the best performing method, CGCS, our method outperforms it by 2.33% and 1.26% on 1-shot and 5-shot classification tasks, respectively. (2) On the CIFAR-FS dataset, our proposed method also achieves the best performance. Compared with the best performing method, PSST, our method outperforms it by 2.67% and 0.42% on 1-shot and 5-shot classification tasks, respectively. (3) On the CUB-200-2011 dataset, our proposed method achieves the best classification performance. Compared with the best performing method, HGNN, our method outperforms it by 1.42% and 0.99% on 1-shot and 5-shot classification tasks, respectively.

4.2. Electrical images Dataset

4.2.1. EEI-100 Dataset

EEI-100(electric equipment image of 100 classes) is a few shot electric equipment classification dataset established by us, which contains 100 classes of electric equipment with 4000 images. The majority of the images were obtained through on-site collection, with a small number of images sourced from online platforms. To the best of our knowledge, this is one of the first datasets specifically designed for classification of electrical equipment. This dataset is an extension of our previous EEI-40 [29]. It includes substation equipment, distribution station equipment and common electrical equipment, ranging from large-scale equipment such as heavy-duty transformers to small-scale equipment such as circuit breakers. A few images from the proposed dataset illustrated in Figure 3. More images illustrated in Appendix A.

4.2.2. Parametric Analysis Experiment

By following the approach outlined in Section 4.1.4, the values of parameters α 1 and α 2 are determined to optimize the performance of the model on EEI-100. Experimental results show that α 1 has the optimal value of 0.6 within the range of [0.1,1], as illustrated in Figure 4(a). Similarly, α 2 has the optimal value of 0.1 within the range of [0.01,0.1], as illustrated in Figure 4(b).

4.2.3. Comparison Experiment with Existing Methods

To demonstrate the superiority of our proposed method in the classification of electric equipment images, this section presents a comparative experiment with three existing methods, namely CGCS, Neg-Cosine, and HGNN, on the EEI-100 dataset. These three methods have recently achieved good performance on public datasets. The classification accuracy of the test set is presented in Table 3. Specifically, our method achieves the highest classification accuracy (up to 94.12%) compared to the other methods.

5. Conclusions

In conclusion, this paper proposes a novel few-shot electrical image classification algorithm based on knowledge distillation. The proposed algorithm addresses the challenges of limited storage capacity of intelligent mobile devices and the difficulty in obtaining large-scale labeled data for all categories of electrical equipment. By adopting few-shot learning method and global and local knowledge distillation, the proposed algorithm achieves high classification accuracy with only a small number of samples. The results on EEI-100 demonstrate that the prediction accuracy of the proposed method reaches 94.12% by collecting only 5-shot images. This research provides a promising solution for the online inspection of electrical equipment, which can significantly enhance the efficiency of detection and maintenance in the power system. Future work can explore the application of the proposed algorithm in other domains and investigate the potential of combining it with other deep learning techniques to further improve the performance.

Author Contributions

Conceptualization, B.Z. and J.G.; methodology, B.Z.; software, C.Y.; validation, B.Z. and J.Z.; formal analysis, B.Z.; investigation, X.Z.; resources, J.Z.; data curation, X.Z.; writing—original draft preparation, B.Z.; writing—review and editing, J.G.

Funding

This research was supported in part by the National Natural Science Foundation of China under Grant No. U2066203 and No. 61973178, and in part by the Key Research & Development Program of Jiangsu Province under Grant No. BE2021063.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset proposed in this paper can be obtained from the author with a reasonable request.

Conflicts of Interest

The authors declare no conflict of interest in preparing this article.

Abbreviations

UAV Unmanned aerial vehicle
CNN convolutional neural network
FSL Few-shot learning
FSC Few-shot classification

Appendix A

In this appendix, more images of the EEI-100 dataset proposed by us are presented. However, we regret to inform that due to the fact that some image data were collected in specific scenarios, the device information in the pictures cannot be disclosed. Therefore, we are unable to fully release the entire dataset here.
Figure A1. Few images of EEI-100 dataset.
Figure A1. Few images of EEI-100 dataset.
Preprints 72437 g0a1

References

  1. Peng, J.; Sun, L.; Wang, K.; Song, L. ED-YOLO power inspection UAV obstacle avoidance target detection algorithm based on model compression. Chinese Journal of Scientific Instrument 2021, 10, 161–170. [Google Scholar]
  2. Geoffrey, H.; Oriol, V.; Jeff, D. Distilling the Knowledge in a Neural Network. arXiv:1503.02531, Mar 2015. arXiv:1503.02531, Mar 2015.
  3. Adriana, R.; Nicolas, B.; Samira, E.K.; Antoine, C.; Carlo, G.; Yoshua, B. FitNets: Hints for Thin Deep Nets. Dec 2014. arXiv:1412.6550.
  4. Bogdan, T.N.; Bruno, M.; Rafael, W.; Victor, B.G.; Vanderlei, Z.; Lourival, L. A Computer Vision System for Monitoring Disconnect Switches in Distribution Substations. IEEE Transactions on Power Delivery 2022, 37, 833–841. [Google Scholar] [CrossRef]
  5. Zhang, Z.D.; Zhang, B.; Lan, Z.C.; Lu, H.C.; Li, D.Y.; Pei, L.; Yu, W.X. FINet: An Insulator Dataset and Detection Benchmark Based on Synthetic Fog and Improved YOLOv5. IEEE Transactions on Instrumentation and Measurement 2022, 71, 1–8. [Google Scholar] [CrossRef]
  6. Xu, Y.; Li, Y.; Wang, Y.; Zhong, D.; Zhang, G. Improved few-shot learning method for transformer fault diagnosis based on approximation space and belief functions. Expert Systems with Applications 2021, 167, 114105. [Google Scholar] [CrossRef]
  7. Yi, Y.; Chen, Z.; Wang, L. Intelligent Aging Diagnosis of Conductor in Smart Grid Using Label-Distribution Deep Convolutional Neural Networks. IEEE Transactions on Instrumentation and Measurement 2022, 71, 1–8. [Google Scholar] [CrossRef]
  8. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep network. Proceedings of the 34th International Conference on Machine Learning (ICML), Sydney, Australia, 7 Aug 2017.
  9. Li, Z.; Zhou, F.; Chen, F.; Li, H. Meta-SGD: Learning to Learn Quickly for Few-Shot Learning. Jul 2017. arXiv:1707.09835.
  10. Ravi, S.; Larochelle, H. Optimization as a model for few-shot learning. 5th International Conference on Learning Representations (ICLR), Toulon, France, 24 Apr 2017.
  11. Wu, Z.; Li, Y.; Guo, L.; Jia, K. PARN. Position-Aware Relation Networks for few-shot learning. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 Oct 2019.
  12. Gidaris, S.; Bursuc, A.; Komodakis, N.; Perez, P.; Cord, M.; Ecole, L. Boosting few-shot visual learning with self-supervision. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 Oct 2019.
  13. Zhang, H.; Zhang, J.; Koniusz, P. Few-shot learning via saliency-guided hallucination of samples. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), California, USA, 16 June 2019.
  14. Hou, R.; Chang, H.; Ma, B.; Shan, S.; Chen, X. Cross attention network for few-shot classification. Proceedings of the 33rd International Conference on Neural Information Processing Systems (NIPS), Vancouver, Canada, 8 Dec 2019.
  15. Guo, Y.; Cheung, N. Attentive weights generation for few shot learning via information maximization. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, USA, 13 June 2020.
  16. Li, H.; Eigen, D.; Dodge, S.; Zeiler, M.; Wang, X. Finding task-relevant features for few-shot learning by category traversal. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), California, USA, 16 June 2019.
  17. Nguyen, V.N.; Løkse, S.; Wickstrøm, K.; Kampffmeyer, M.; Roverso, D.; Jenssen, R. SEN: a novel feature normalization dissimilarity measure for prototypical few-Shot learning networks. Proceedings of the 16th European conference on computer vision (ECCV), Glasgow, Scotland, 23 Aug 2020.
  18. Wertheime, D.; Tang, L.; Hariharan, B. Few-shot classification with feature map reconstruction networks. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), online, 19 June 2021.
  19. Li, W.; Wang, L.; Xu, J.; Huo, J.; Gao, Y.; Luo, J. Revisiting local descriptor based image-to-class measure for few-shot learning. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), California, USA, 16 June 2019.
  20. Zhang, C.; Cai, Y.; Lin, G.; Shen, C. DeepEMD: few-shot image classification with differentiable Earth Mover’s distance and structured classifiers. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, USA, 13 June 2020.
  21. Chen, Y.; Liu, Y. ; Kira, Zsolt.; Wang, Y.F.; Huang, J. A closer look at few-shot classification. Proceedings of the 7th International Conference on Learning Representations (ICLR), New Orleans, USA, 6 May 2019.
  22. Liu, B.; Cao, Y.; Lin, Y.; Zhang, Z.; Long, M.; Hu, H. Negative margin matters: understanding margin in few-shot classification. Proceedings of the 16th European conference on computer vision (ECCV), Glasgow, Scotland, 23 Aug 2020.
  23. Mangla, P.; Singh, M.; Sinha, A.; Kumari, N.; Balasubramanian, V.; Krishnamurthy, B. Charting the right manifold: Manifold mixup for few-shot learning. The IEEE Winter Conference on Applications of Computer Vision (WACV), Hawaii, USA, 7 Janu 2020.
  24. Su, J.; Maji, S.; Hariharan, B. When does self-supervision improve few-shot learning. Proceedings of the 16th European conference on computer vision (ECCV), Glasgow, Scotland, 23 Aug 2020.
  25. Shao, S.; Xing, L.; Wang, Y.; Xu, R.; Zhao, C.; Wang, Y.J.; Liu, B. MHFC: multi-head feature collaboration for few-shot learning. Proceedings of the 29th ACM International Conference on Multimedia (MM), Online, 20 Otc 2021.
  26. Zagoruyko, S.; Komodakis, N. Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. 5th International Conference on Learning Representations (ICLR), Toulon, France, 24 Apr 2017.
  27. Park, W.; Kim, D.; Lu, Y.; Cho, M. Relational knowledge distillation. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), California, USA, 16 June 2019.
  28. Peng, B.; Jin, X. ; Liu, J; Zhou, S.; Wu, Y.; Liu, Y.; Li, D.; Zhang, Z. Correlation congruence for knowledge distillation. Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27 Oct 2019.
  29. Zhou, B.; Zhang, X. ; Zhao, J; Zhao, F.; Yan, C.; Xu, Y.; Gu, J. Few-shot electric equipment classification via mutual learning of transfer-learning model. IEEE 5th International Electrical and Energy Conference (CIEEC), Nangjing, China, 27 May 2022.
Figure 1. Network architecture.
Figure 1. Network architecture.
Preprints 72437 g001
Figure 2. Test experiments under different values of α 1 and α 2 and α 2 on three public datasets. (a) 1-shot test accuracy under different values of α 1 ; (b) 5-shot test accuracy under different values of α 1 ; (c) 1-shot test accuracy under different values of α 2 ; (d) 5-shot test accuracy under different values of α 2 .
Figure 2. Test experiments under different values of α 1 and α 2 and α 2 on three public datasets. (a) 1-shot test accuracy under different values of α 1 ; (b) 5-shot test accuracy under different values of α 1 ; (c) 1-shot test accuracy under different values of α 2 ; (d) 5-shot test accuracy under different values of α 2 .
Preprints 72437 g002
Figure 3. Some images of EEI-100 dataset. They represent different electrical equipment. (a)Wind power tower; (b)Heavy-duty transformer; (c)Heavy-duty distribution cabinet; (d)Energy storage battery pack; (e)Electrical insulator; (f) Split-core current transformer; (g) Three-phase moto; (h) Heavy-duty circuit breaker; (i) Contactor; (j) Cooling fan; (k) Electric energy meter; (l)Dragline board.
Figure 3. Some images of EEI-100 dataset. They represent different electrical equipment. (a)Wind power tower; (b)Heavy-duty transformer; (c)Heavy-duty distribution cabinet; (d)Energy storage battery pack; (e)Electrical insulator; (f) Split-core current transformer; (g) Three-phase moto; (h) Heavy-duty circuit breaker; (i) Contactor; (j) Cooling fan; (k) Electric energy meter; (l)Dragline board.
Preprints 72437 g003
Figure 4. Test experiments under different values of α 1 and α 2 on EEI-100 dataset. (a) 1-shot and 5-shot test accuracy under different values of α 1 ; (b) 1-shot and 5-shot test accuracy under different values of α 2 .
Figure 4. Test experiments under different values of α 1 and α 2 on EEI-100 dataset. (a) 1-shot and 5-shot test accuracy under different values of α 1 ; (b) 1-shot and 5-shot test accuracy under different values of α 2 .
Preprints 72437 g004
Table 1. Results(%) of ablation experiment.
Table 1. Results(%) of ablation experiment.
Method Backbone MiniImageNet CIFAR-FS CUB
1-shot 5-shot 1-shot 5-shot 1-shot 5-shot
Global Conv4 57.32±0.84 72.90±0.64 66.40±0.93 80.44±0.67 70.20±0.93 83.88±0.57
Local Conv4 57.65±0.83 73.06±0.64 66.63±0.93 80.64±0.67 70.12±0.93 83.66±0.57
Global-Local Conv4 57.86±0.83 73.38±0.62 67.04±0.91 80.84±0.68 70.44±0.92 84.19±0.56
Table 2. Results(%) of comparison experiment.
Table 2. Results(%) of comparison experiment.
Method Backbone MiniImageNet CIFAR-FS CUB
1-shot 5-shot 1-shot 5-shot 1-shot 5-shot
Meta-learning
Relational Conv4 50.44±0.82 65.32±0.70 55.00±1.00 69.30±0.80 62.45± 0.98 76.11± 0.69
MetaOptSVM Conv4 52.87±0.57 68.76±0.48 - - - -
PN+rot Conv4 53.63±0.43 71.70±0.36 - - - -
CovaMNet Conv4 51.19±0.76 67.65± 0.63 - - 52.42±0.76 63.76±0.64
DN4 Conv4 51.24±0.74 71.02±0.64 - - 46.84±0.81 74.92±0.64
MeTAL Conv4 52.63±0.37 70.52±0.29 - -
HGNN Conv4 55.63±0.20 72.48±0.16 - - 69.02±0.22 83.20±0.15
DSFN Conv4 50.21±0.64 72.20±0.51 - - - -
PSST Conv4 - - 64.37±0.33 80.42± 0.32 - -
Transfer-learning
Baseline++ Conv4 48.24±0.75 66.43±0.63 - - 60.53±0.83 79.34±0.61
Neg-Cosine Conv4 52.84±0.76 70.41±0.66 - - - -
SKD Conv4 48.14 66.36 - - - -
CGCS Conv4 55.53±0.20 72.12±0.16 - - - -
Our method Conv4 57.86±0.83 73.38±0.62 67.04±0.91 80.84±0.68 70.44±0.92 84.19±0.56
Table 3. Experiment result (%) on EEI-100 dataset.
Table 3. Experiment result (%) on EEI-100 dataset.
Method 1-shot 5-shot
CGCS 72.85±0.68 89.68±0.27
Neg-Cosine 74.57±0.63 90.54±0.25
HGNN 75.61±0.62 93.54±0.24
Our method 75.80±0.67 94.12±0.20
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.

Downloads

141

Views

30

Comments

0

Subscription

Notify me about updates to this article or when a peer-reviewed version is published.

Email

Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2025 MDPI (Basel, Switzerland) unless otherwise stated