Valero-Mas, Jose J., Gallego, Antonio-Javier, Rico-Juan, Juan Ramón An overview of ensemble and feature learning in few-shot image classification using siamese networks Multimedia Tools and Applications. 2024, 83: 19929-19952. https://doi.org/10.1007/s11042-023-15607-3 URI: http://hdl.handle.net/10045/136763 DOI: 10.1007/s11042-023-15607-3 ISSN: 1380-7501 (Print) Abstract: Siamese Neural Networks (SNNs) constitute one of the most representative approaches for addressing Few-Shot Image Classification. These schemes comprise a set of Convolutional Neural Network (CNN) models whose weights are shared across the network, which results in fewer parameters to train and less tendency to overfit. This fact eventually leads to better convergence capabilities than standard neural models when considering scarce amounts of data. Based on a contrastive principle, the SNN scheme jointly trains these inner CNN models to map the input image data to an embedded representation that may be later exploited for the recognition process. However, in spite of their extensive use in the related literature, the representation capabilities of SNN schemes have neither been thoroughly assessed nor combined with other strategies for boosting their classification performance. Within this context, this work experimentally studies the capabilities of SNN architectures for obtaining a suitable embedded representation in scenarios with a severe data scarcity, assesses the use of train data augmentation for improving the feature learning process, introduces the use of transfer learning techniques for further exploiting the embedded representations obtained by the model, and uses test data augmentation for boosting the performance capabilities of the SNN scheme by mimicking an ensemble learning process. The results obtained with different image corpora report that the combination of the commented techniques achieves classification rates ranging from 69% to 78% with just 5 to 20 prototypes per class whereas the CNN baseline considered is unable to converge. Furthermore, upon the convergence of the baseline model with the sufficient amount of data, still the adequate use of the studied techniques improves the accuracy in figures from 4% to 9%. Keywords:Few-shot learning, Siamese networks, Data Augmentation, Transfer learning Springer Nature info:eu-repo/semantics/article