TY - JOUR
T1 - Convolutional neural network models with low spatial variability hamper the transfer learning process
AU - Bravo-Diaz, Alejandra
AU - Moreno, Sebastián
AU - Lopatin, Javier
N1 - Publisher Copyright:
© The Author(s), under exclusive licence to Springer-Verlag London Ltd., part of Springer Nature 2025.
PY - 2025/7
Y1 - 2025/7
N2 - Using pre-trained convolutional neural networks (CNNs) architectures have proven effective in high-resolution remote sensing, even when homogeneous and few data samples are used. However, it is still uncertain how well models trained with limited spatial information can transfer the learning process from one domain to a new domain (transductive transfer learning). This paper evaluates transductive transfer learning in CNN regression models using RGB-based data captured by unoccupied aerial vehicles on five sites, using the prediction of Pinus radiata canopy coverage as a case study. We trained five models, one per site, analyzing their internal performance using fine-tuning and feature extraction training approaches. Then, we evaluated their transfer learning ability to new unseeing sites. We found that the trained models perform accurately within their domain, as previous research demonstrates (R2 > 0.90 using fine-tuning). However, we depicted varying performances during transfer learning, with R2 ranging from − 0.68 to 0.63 for feature extraction and from − 4.42 to 0.77 for fine-tuning. Our results show that these poor performances are independent of the training approach (fine-tuning or feature extraction), the number of observations, or the complexity of the model. In contrast, the success during transfer learning is closely linked to the similarity between the source and target domains, which is often unknown when predicting new data. These results depict the importance of carefully planning the future use of such models for their sustainability and generalization over time.
AB - Using pre-trained convolutional neural networks (CNNs) architectures have proven effective in high-resolution remote sensing, even when homogeneous and few data samples are used. However, it is still uncertain how well models trained with limited spatial information can transfer the learning process from one domain to a new domain (transductive transfer learning). This paper evaluates transductive transfer learning in CNN regression models using RGB-based data captured by unoccupied aerial vehicles on five sites, using the prediction of Pinus radiata canopy coverage as a case study. We trained five models, one per site, analyzing their internal performance using fine-tuning and feature extraction training approaches. Then, we evaluated their transfer learning ability to new unseeing sites. We found that the trained models perform accurately within their domain, as previous research demonstrates (R2 > 0.90 using fine-tuning). However, we depicted varying performances during transfer learning, with R2 ranging from − 0.68 to 0.63 for feature extraction and from − 4.42 to 0.77 for fine-tuning. Our results show that these poor performances are independent of the training approach (fine-tuning or feature extraction), the number of observations, or the complexity of the model. In contrast, the success during transfer learning is closely linked to the similarity between the source and target domains, which is often unknown when predicting new data. These results depict the importance of carefully planning the future use of such models for their sustainability and generalization over time.
KW - Domains
KW - Feature extraction
KW - Fine-tuning
KW - Pre-trained model
KW - Small data
KW - Transductive transfer learning
UR - https://www.scopus.com/pages/publications/105004320336
U2 - 10.1007/s00521-025-11267-6
DO - 10.1007/s00521-025-11267-6
M3 - Article
AN - SCOPUS:105004320336
SN - 0941-0643
VL - 37
SP - 13927
EP - 13942
JO - Neural Computing and Applications
JF - Neural Computing and Applications
IS - 19
ER -