From Few Images to High Accuracy: Augmentation and Embedding Methods for Date Fruit Ripeness

Document Type : Research Article

Authors

1 Dept. of Biosystem engineering, University of Mohaghegh Ardabili

2 2Department of Computer Engineering, Sharif University of Technology, Tehran 14588-89694, Iran.

3 3Department of Biosystems Engineering, Gorgan University of Agricultural Sciences and Natural Resources, Gorgan, Iran.

Abstract

Manual date harvesting and sorting remain labor-intensive and error-prone, particularly when distinguishing intermediate ripeness stages such as Rotab. We present an image-based classification pipeline for the Berhi cultivar that assigns fruit to three ripeness stages—Khalal, Rotab, and Tamar—using compact deep structures and training strategies suited to small datasets. Rather than relying on generative or adversarial methods, our approach emphasizes (i) careful augmentation (classical transforms, automated policies, and sample-mixing), (ii) transfer and self-supervised pre training, and (iii) embedding- and metric-learning alternatives, with ensembles and test-time augmentation used as optional accuracy/robustness boosters. On a 150-image dataset (50 images per class) evaluated with 5-fold cross-validation, a ResNet18 baseline reaches about 95% average accuracy. Automated augmentation combined with MixUp/CutMix improves accuracy to 97%, and self-supervised pre training plus advanced augmentation and ensembling attain peak performance near 98%. Improvements are most pronounced for the visually ambiguous Rotab class. We also report practical robustness measures (common corruptions, geometric stability, and calibration), which show that augmentation and pre training substantially increase stability under realistic input variability. These results indicate that, for small and visually subtle datasets, augmentation and pre training—rather than synthetic data generation—offer a pragmatic path to high accuracy and robust behavior.

Keywords

Main Subjects



Articles in Press, Accepted Manuscript
Available Online from 22 November 2025
  • Receive Date: 26 October 2025
  • Revise Date: 13 November 2025
  • Accept Date: 22 November 2025
  • First Publish Date: 22 November 2025
  • Publish Date: 22 November 2025