Submitted:
18 July 2025
Posted:
21 July 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Materials and Methods
2.1. Fish Detection Model and Performance Metrics
2.2. Image Data
2.3. Dataset
2.4. Model Training and Validation
2.5. Inference and Inference Testing
2.6. Estimation of Actual Fish Counts
3. Results
3.1. Training and Validation
3.2. Inference and Inference Testing
3.3. Estimated Number of Individuals
4. Discussion
4.1. Comparison of Validation and Inference Performance
4.2. Implications of Estimated Individual Counts and Detection Performance
5. Conclusions
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Maeda, S.; Yoshida, K.; Kuroda, H. Turbulence and energetics of fish nest and pool structures in agricultural canal. Paddy and Water Environment 2018, 16, 493–505. [Google Scholar] [CrossRef]
- Takahara, T.; Minamoto, T.; Yamanaka, H.; Doi, H.; Kawabata, Z. Estimation of fish biomass using environmental DNA. Plos One 2012, 7, e35868. [Google Scholar] [CrossRef] [PubMed]
- Jalal, A.; Salman, A.; Mian, A.; Shortis, M.; Shafait, F. Fish detection and species classification in underwater environments using deep learning with temporal information. Ecological Informatics 2020, 57, 101088. [Google Scholar] [CrossRef]
- Kandimalla, V.; Richard, M.; Smith, F.; Quirion, J.; Torgo, L.; Whidden, C. Automated detection, classification and counting of fish in fish passages with deep learning. Frontiers in Marine Science 2022, 8, 823173. [Google Scholar] [CrossRef]
- Knausgård, K.M.; Wiklund, A.; Sørdalen, T.K.; Halvorsen, K.T.; Kleiven, A.R.; Jiao, L.; Goodwin, M. Temperate fish detection and classification: a deep learning based approach. Applied Intelligence 2022, 52, 6988–7001. [Google Scholar] [CrossRef]
- Muksit, A.A.; Hasan, F.; Emon, M.F.H.B.; Haque, M.R.; Anwary, A.R.; Shatabda, S. YOLO-Fish: A robust fish detection model to detect fish in realistic underwater environment. Ecological Informatics 2022, 72, 101847. [Google Scholar] [CrossRef]
- Zhang, Z.; Qu, Y.; Wang, T.; Rao, Y.; Jiang, D.; Li, S.; Wang, Y. An improved YOLOv8n used for fish detection in natural water environments. Animals 2024, 14, 2022. [Google Scholar] [CrossRef] [PubMed]
- Yang, J.; Takani, A.; Yanaba, N.; Abe, S.; Hayashi, Y.; Kiino, A. Research on automated recognition and counting of fish species by image analysis using deep learning. Transactions of AI and Data Science 2024, 5, 13–21. (In Japanese) [Google Scholar]
- Pan, S.; Furutani, R.; Yoshida, K.; Yamashita, Y.; Kojima, T.; Shiraga, Y. A study on detection of fishways running-up juvenile ayu using underwater camera images and deep learning. Journal of Japan Society of Civil Engineers, Ser. B1 (Hydraulic Engineering) 2022, 78, I_127–I_132. (In Japanese) [Google Scholar] [CrossRef] [PubMed]
- Takeda, K.; Yoshikawa, N.; Miyazu, S. Development of an automatic fish counting method in ultrasonic echo images using deep learning. In Proceedings of the JSIDRE (Japanese Society of Irrigation, Drainage, and Rural Engineering) Annual Congress; (In Japanese). 2023; pp. 6–33. [Google Scholar]
- Zhang, Z.; Li, J.; Su, C.; Wang, Z.; Li, Y.; Li, D.; Chen, Y.; Liu, C. A method for counting fish based on improved YOLOv8. Aquacultural Engineering 2024, 107, 102450. [Google Scholar] [CrossRef]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 779–788. 2016. [Google Scholar]
- Maeda, S.; Takagi, S.; Yoshida, K.; Kuroda, H. Spatiotemporal variation of sedimentation in an agricultural drainage canal with eco-friendly physical structures: a case study. Paddy and Water Environment 2021, 19, 189–198. [Google Scholar] [CrossRef]
- Maeda, S.; Lin, X.; Kuroda, H. Evaluation of the impact of fish nest size on the discharge rate of filamentous algae. Journal of Japan Society of Civil Engineers 2025, 81, 24–16205. (In Japanese) [Google Scholar] [CrossRef]
- Tzutalin. LabelImg. Git code. Available online: https://github.com/heartexlabs/labelImg (accessed on 9 July 2025).









| Category | Training data | Validation data |
| Number of images (total) | 546 images | 129 images |
| Number of background-only images | 57 images | 14 images |
| Number of images with fish | 489 images | 115 images |
| Pseudorasbora parva | 305 individuals | 76 individuals |
| Cyprinus carpio | 333 individuals | 74 individuals |
| Configuration | Parameter |
| GPU | NVIDIA Quadoro P4000 |
| Operating System | Windows 10 Pro, 64-bit |
| Python | 3.8.5 |
| PyTorch | 2.2.0+cu118 |
| CUDA | 11.8 |
| Optimizer | AdamW |
| Initial learning rate | 0.001 |
| β1 | 0.937 |
| β2 | 0.999 |
| Weight decay | 0.0005 |
| Learning rate scheduler | Linear decay |
| Epochs | 500 |
| Input Image Resolution | 960 × 960 pixels |
| Batch Size | 16 |
| Images | Instances | Precision(%) | Recall(%) | F1 score(%) | mAP50(%) | mAP50-95(%) | |
|---|---|---|---|---|---|---|---|
| All species | 129 | 150 | 95.1 | 88.4 | 91.6 | 95.5 | 69.0 |
| Pseudorasbora parva | 129 | 76 | 95.2 | 78.1 | 85.8 | 91.6 | 58.8 |
|
Cyprinus carpio |
129 | 74 | 95.0 | 98.6 | 96.8 | 99.4 | 79.1 |
| Images | Instances | Precision(%) | Recall(%) | F1 score(%) | mAP50(%) | mAP50-95(%) | |
|---|---|---|---|---|---|---|---|
| All species | 300 | 389 | 88.9 | 47.1 | 61.6 | 67.8 | 47.0 |
| Pseudorasbora parva | 300 | 375 | 96.5 | 37.0 | 53.5 | 72.0 | 48.7 |
|
Cyprinus carpio |
300 | 14 | 81.2 | 57.1 | 67.1 | 63.6 | 45.3 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).