Banner

Indian Journal of Animal Research

  • Chief EditorM. R. Saseendranath

  • Print ISSN 0367-6722

  • Online ISSN 0976-0555

  • NAAS Rating 6.40

  • SJR 0.233, CiteScore: 0.606

  • Impact Factor 0.5 (2025)

Frequency :
Monthly (January, February, March, April, May, June, July, August, September, October, November and December)
Indexing Services :
Science Citation Index Expanded, BIOSIS Preview, ISI Citation Index, Biological Abstracts, Scopus, AGRICOLA, Google Scholar, CrossRef, CAB Abstracting Journals, Chemical Abstracts, Indian Science Abstracts, EBSCO Indexing Services, Index Copernicus

A Deep Convolutional Neural Network for Multi-class Fish Disease Classification using VGGNet-19 Architecture

Kyung-Won Cho1, Taeho Kim2, In Seop NA3,*
1Cybertech Co., 276 Greenro, Educastle apartment 102, Unit 1705 Naju-si, Jeollanam-do, Republic of Korea.
2Department of Marine Production Management and Smart Aquaculture Research Center, Chonnam National University, 50, Daehak-ro, Yeosu-si, Jeollanam-do, 59626, Republic of Korea.
3Division of Culture Contents, Chonnam National University, 50, Daehak-ro, Yeosu-si, Jeollanam-do, 59626, Republic of Korea.

Background: Accurate fish disease identification is important for ecological research, biodiversity monitoring and conservation. Traditional methods are time-consuming and error-prone. Deep learning provides an automated solution for species classification using image data. Fish disease classification using image-based recognition is an emerging research area, critical for the aquaculture industry. Accurate disease detection is essential for preventing outbreaks, minimizing economic losses and ensuring fish health.

Methods: This study employs the VGGNet-19 architecture to detect fish diseases. The dataset, sourced from Kaggle, contains images of fish affected by various diseases. Images are pre-processed with resizing, normalization and data augmentation techniques, including random flipping, rotation and zoom. The model is trained with an 80%-10%-10% split for training, validation and testing. Performance is evaluated using accuracy, a confusion matrix, a classification report and ROC curves.

Result: The custom model achieved an overall accuracy of 90.96%. It performed well across most disease categories, correctly distinguishing between healthy and infected fish. Some misclassifications were observed between similar diseases, indicating areas for improvement. Despite these challenges, the results demonstrate the effectiveness of deep learning for fish disease classification.

In recent years, global aquaculture production has witnessed significant growth, reaching a record 130.9 million tonnes in 2022, with 94.4 million tonnes consisting of aquatic animals (FAO, 2023). Asia continues to dominate the sector, contributing 91.4% of total production, followed by the Americas (3.5%), Europe (2.8%), Africa (1.9%) and Oceania (0.2%) (WAS, 2023). Furthermore, global fish production increased to 186.6 million metric tons in 2023, reflecting the rising demand for fish and seafood products worldwide (Statista, 2024). This growth highlights the increasing need for effective fish disease detection and management strategies to sustain industry expansion and reduce economic losses.
       
Fish farming contributes an important part of global food production and provides a substantial source of protein to millions of people worldwide. Due to the rising global demand for protein, fish farming has seen rapid growth. Fish consumption is a vital dietary source of protein, supplying essential nutrients to billions of people globally. In fact, global fish consumption per capita has increased from 9.0 kg in 1961 to 20.2 kg in 2015 and further up to 22.8 kg by 2020 (Coppola et al., 2021). This surge has led to significant advances in aquaculture technologies, enabling increased fish production to meet the growing demand. In 2020, the total annual production of fish and seafood was around 200 million tonnes, including both wild fish catch and farmed fish (FAO, 2020). Over the past two decades, aquaculture production has far outpaced the capture fisheries sector, further emphasizing the sector’s importance in global food security (Datasource - OurWorldInData.org/fish-and-overfishing).
       
However, maintaining the health of farmed fish remains a complex and challenging task. Fish diseases can lead to large-scale losses. According to estimates, 20-25% of global fish production is wasted due to factors such as disease outbreaks, delayed transportation and other industrial obstructions (Begum et al., 2024; Zhang et al., 2024). Diseases in fish are particularly concerning in farmed environments, as the confined and contained nature of aquaculture facilitates the rapid spread of pathogens, much faster than in wild fish populations (Sven et al., 2021).
       
Traditional methods are time-consuming and prone to human error. As fish farming expands, there is an increasing need for more efficient and automated methods of disease detection and classification. One promising method is the application of artificial intelligence (AI), particularly deep learning techniques such as Convolutional Neural Networks (CNNs). CNNs have demonstrated exceptional performance in various image-based classification tasks, including medical diagnostics and plant disease recognition. Despite their success in other fields, the application of CNNs to fish disease classification has not been extensively explored, presenting an opportunity to improve fish health monitoring systems (AlZubi, 2023, Bagga et al., 2024; Cho, 2024; Hai and Duong, 2024; Maltare et al., 2023).
       
Islam et al. (2024) reviewed modern technologies for managing fish diseases in aquaculture. They highlighted the critical need for early detection of infectious agents such as bacteria and parasites. High-density aquaculture settings accelerated disease spread and could cause up to 100% mortality if untreated. Technologies like IoT sensors, AI and machine learning offer proactive monitoring and early pathogen detection. This approach allowed farmers to address potential outbreaks before they occurred. Vasumathi et al., (2024) highlighted the efforts of the Fish4Knowledge project, which created a large library of fish images and videos for training CNN models. A major challenge in this development was the wide variety of fish appearances. These variations depended on factors such as species, age and health. The differences in size, shape, color and pattern made it difficult for AI algorithms to accurately distinguish between healthy and diseased fish. Jiang et al. (2024) conducted a study on fish disease detection using big data and deep learning. The researchers collected and labeled data and then extracted key features related to fish diseases. They improved the traditional detection model by incorporating the Akconv network structure and the SimAM attention module into the YOLOv6 model. Experiments showed that the enhanced YOLOv6 model outperformed the original, with an increase in its mean average precision (mAP) by 1.8 percentage points.
       
Isaac et al. (2024) proposed five deep learning models (CNN, DenseNet121, ViT, VGG16 and VGG19) for fish disease detection, evaluating their performance using accuracy. The dataset underwent preprocessing, including resizing, normalization and augmentation, to balance classes. The models were trained and evaluated, with deep transfer learning applied to improve generalization. Nguyen et al. (2024) used MobileNet models to automate fish disease classification, aiming to reduce costs and time in aquaculture. The models were trained on the SalmonScan dataset, with MobileNetV1 achieving accuracies of 96.30% and 95.47% with and without data augmentation. The study showed that MobileNet could autonomously learn disease features, eliminating the need for manual feature extraction and offered more efficient disease detection compared to traditional methods.
               
This paper proposes a solution to the problem by applying a custom version of the VGGNet-19 architecture to classify fish diseases from image data. VGGNet-19, originally developed for large-scale image classification tasks, has proven to be highly effective in learning rich hierarchical features from images. By customizing and fine-tuning this architecture for the specific application, high classification accuracy is achieved on a diverse dataset of fish images that can improve fish health management systems.
The model is developed in an Anaconda environment using Jupyter Notebook, with training conducted on a system with 16 GB RAM, Python v3.11 and a TPU-v3:8 setup. The TPU is optimized for machine learning tasks and is supported by TensorFlow. Keras, a high-level neural network API, operates on top of TensorFlow, enabling efficient computation with the 8-core TPU for faster machine learning processing.
 
Dataset collection
 
This study utilizes an open-source dataset to develop a fish disease detection model using VGGNet-19-CNN. The dataset includes images of fish affected by seven conditions, categorized into bacterial (Aeromoniasis, Bacterial Gill Disease, Bacterial Red Disease), fungal (Saprolegniasis), parasitic (Parasitic Disease) and viral (White Spot Disease) infections, along with a healthy fish class. Each category contains 250 images, ensuring a balanced dataset for training and evaluation. The dataset was obtained from a publicly available repository on Kaggle (https://www.kaggle.com/datasets/subirbiswas19/freshwater-fish-disease-aquaculture-in-south-asia?select=SB-FishDisease) and collected from various sources, including a university agricultural department, an agricultural farm in Odisha, India and agricultural website portals. To ensure effective model training and evaluation, the dataset was divided into three subsets: 80% for training, 10% for validation and 10% for testing. This division was performed using the train-test-split function from the sklearn model selection library, ensuring a fair and randomized distribution across all categories.
 
Data preprocessing and augmentation
 
To improve model generalization and avoid overfitting, several preprocessing and augmentation techniques were applied to the dataset. First, image resizing was done to ensure all images had the same dimensions. The VGGNet-19 model requires input images of size 224 × 224 pixels, so all images in the dataset were resized to match this requirement. This step ensures that the model receives uniform input, which is essential for efficient training. Next, pixel normalization was applied. The pixel values of each image were scaled to a range of [0, 1]. To further enhance the dataset, data augmentation techniques were used. These techniques artificially increase the size of the training set and make the model more robust. Random rotation was applied to simulate different orientations of the fish. This helps the model recognize diseases from various angles. Horizontal flipping was also used to create mirror images, which accounts for variations in fish orientation. Additionally, random zoom was used to simulate varying distances, allowing the model to focus on different areas of the fish. Random shearing and shifting were applied to introduce slight distortions, increasing the diversity of the training images. These augmentation techniques helped prevent overfitting by providing the model with more varied data and improved its ability to generalize to new, unseen images.
 
VGGNet-19 Architecture
 
VGGNet-19 is a deep convolutional neural network (CNN) known for its simplicity and effectiveness in image classification. It consists of 19 layers, including 16 convolutional layers, 3 fully connected layers and 5 max-pooling layers (Fig 1). VGGNet-19 uses 3×3 convolutional filters, which allow the model to capture fine-grained features in the images. The hyperparameters used for modeling are summarized in Table 1.

Fig 1: VGGNet-19 modal.



Table 1: summarizes the hyperparameters used in the VGGNet-19 model.


       
The model accepts input images of size 224×224×3. This means the input is 224 pixels in height and width, with 3 color channels (RGB). The network has several convolutional blocks, each containing two or three convolutional layers followed by max-pooling layers. These blocks help the model learn hierarchical features, from basic edges to more complex patterns. The convolution operation can be represented mathematically as
 
Where,
x = Input image.
w = Convolutional kernel.
y i= Resulting feature map.
       
Max-pooling reduces the spatial dimensions of the feature maps. The operation can be written as


After convolutional and pooling layers, the model has two fully connected layers, each with 4096 neurons. These layers help the network make final predictions. The activation function used in the fully connected layers is ReLU


The final layer is a softmax layer with 7 units, one for each disease category. The softmax function converts the network’s output into probabilities

 
Where,
zi = Raw score for class.
i = The softmax function normalizes these scores into a probability distribution.
       
Dropout is applied to the fully connected layers to prevent overfitting. It randomly disables a fraction of neurons during training. The dropout operation can be expressed as

 
The model was trained using the Adam optimizer, which is effective for training deep networks. The learning rate was set to 0.0001 to ensure stable convergence. The Adam optimizer updates parameters using


Where,
θ  = Parameter at time t.
η = Learning rate.
mt = First moment estimate.
νt  = Second moment estimate.
       
The loss function used was Categorical Cross-Entropy. This is appropriate for multi-class classification problems. The formula for categorical cross-entropy is


Where,
C = Number of classes.
yi  = True label for class.
i and wi  = Predicted probability for class i.
       
The model was trained for 100 epochs with early stopping to avoid overfitting. Early stopping halts training if the model’s performance on the validation set does not improve after a set number of epochs. The training process used a batch size of 32 images per iteration. This batch size was chosen to balance training speed and model performance. Model checkpoints were saved during training based on the highest validation accuracy. This ensured that the best-performing model was retained. The model’s performance was evaluated using the following metrics.





In this part, the results obtained from VGGNet19 model are presented in detail. The training and validation metrics were evaluated for 100 epochs and shown in Fig  2. The VGGNet19 model shows improvement over the 100 epochs, with both training and validation losses steadily decreasing. The training loss starts at 1.8844 and drops to 0.0011 by the end, indicating effective learning on the training data. Similarly, the validation loss decreases but stabilizes around 0.1–0.2 after epoch 60, suggesting the model is not generalizing as well to unseen data. Training accuracy steadily increases from 0.3776 to 0.9964, reflecting better performance on the training set. However, validation accuracy plateaus around 0.93–0.95, indicating that the model’s ability to generalize has limitations. The gap between training and validation accuracy points to potential overfitting, where the model fits the training data well but struggles to generalize to the validation set. Despite these fluctuations, the overall trends suggest that with further adjustments, such as regularization, the model’s generalization could improve.

Fig 2: Training and validation outcomes as a function of epochs.


       
The confusion matrix provides useful insights into the model’s performance (Fig 3). It shows both true positives (diagonal values) and misclassifications (off-diagonal values). The model performs well for most diseases. For example, Healthy is correctly classified 24 out of 25 times with only one misclassification to Gill disease. Gill disease is correctly identified in 26 out of 28 instances with two misclassifications. Red disease is correctly classified 25 times, but there are some misclassifications for Aeromoniasis and Saprolegniasis. Aeromoniasis is correctly classified 23 times with minor errors to Saprolegniasis and Healthy.

Fig 3: Confusion matrix.


       
Some diseases are more challenging. White spot disease has three misclassifications, while Parasitic disease is confused with Red disease and Gill disease. These errors likely occur due to similarities in symptoms. The diagonal values are much higher than the off-diagonal values, showing that the model is generally accurate. However, misclassifications still occur, particularly between diseases with overlapping symptoms. The heatmap presented in Fig 3 highlights the model’s strengths and areas that need improvement.
       
The model performance outputs are summarized in Table 2. The VGGNet19 model performs well across most disease categories. It achieves high precision and recall for Healthy, with an F1 score of 0.9410. Aeromoniasis has the highest precision at 0.9580, but its recall is 0.8850, meaning some cases are missed. Gill disease has good recall at 0.9290, but its precision is 0.8670, leading to some false positives. Saprolegniasis performs well with an F1 score of 0.8800, but its precision is slightly lower at 0.8460. Parasitic disease has a balanced precision of 0.9570 and recall of 0.9170, with an F1 score of 0.9360. White spot disease shows strong precision (0.9500), but its recall is lower at 0.8640, resulting in missed cases. The overall accuracy of the model is 90.96%. The macro averages for precision, recall and F1 score are around 0.91. The weighted averages are slightly higher due to class distribution. Overall, the model performs well, but misclassifications still occur. Fine-tuning could improve accuracy, especially for categories like White spot disease.

Table 2: Classification matrices.


       
The ROC AUC scores show different levels of performance across disease categories (Fig 4). The model performs well with Red disease, achieving an AUC of 0.86. Aeromoniasis has an AUC of 0.89, indicating strong classification. Gill disease and Healthy both have excellent AUC scores of 0.91, showing high accuracy. Saprolegniasis has a lower AUC of 0.71, indicating difficulty in distinguishing it from other diseases. Parasitic disease performs reasonably well with an AUC of 0.82. White spot disease has the lowest AUC at 0.67, suggesting the model struggles with this class. Overall, the model is effective for most diseases but needs improvement for Saprolegniasis and White spot disease.

Fig 4: ROC-AUC curve.


       
The results obtained from the VGGNet19 model indicate strong classification performance for fish disease detection, with an overall accuracy of 90.96%. The model effectively learns from training data, as shown by the decreasing loss and increasing accuracy over 100 epochs. However, the observed gap between training and validation accuracy suggests potential overfitting. While the training accuracy reaches 99.64%, the validation accuracy stabilizes around 93-95%, indicating that the model struggles with generalization to unseen data. This highlights the need for additional regularization techniques, such as dropout layers or data augmentation, to enhance robustness.
       
The confusion matrix analysis reveals that most disease classes are correctly classified with high precision and recall. Notably, Healthy samples are accurately identified 96% of the time and Aeromoniasis achieves the highest precision at 95.8%. However, certain disease classes, such as White spot disease and Saprolegniasis, present challenges, as evidenced by their relatively lower recall values of 86.4% and 71%, respectively. These misclassifications likely arise due to visual similarities between disease symptoms, suggesting the need for more discriminative features or the inclusion of additional training samples for these classes.
       
Despite the overall strong classification performance, some diseases are misclassified more frequently than others. White spot disease, for instance, is often confused with Parasitic disease and Gill disease. This may be due to overlapping visual symptoms in infected fish. Similarly, Saprolegniasis exhibits a lower AUC score of 0.71, indicating difficulty in distinguishing it from other infections. The model’s limitations in these areas suggest that additional feature extraction techniques, such as attention mechanisms or ensemble learning, could help improve classification accuracy for the more challenging disease categories.
       
The ROC-AUC analysis further supports these findings, showing that most diseases have AUC scores above 0.80, indicating strong classification ability. However, White spot disease has the lowest AUC (0.67), demonstrating the need for improvement in detecting this particular condition. The imbalance in disease recognition performance suggests that additional data preprocessing, such as synthetic data generation or class-balancing techniques, could help mitigate these issues.
       
Future improvements to the model could involve hyperparameter tuning, employing more advanced architectures such as transformer-based models and leveraging domain-specific augmentation techniques to enhance generalization. Additionally, integrating multi-modal data, such as water quality parameters and behavioral analysis, may further refine disease classification accuracy.
       
The comparison of results obtained in the presented study highlights the importance of the proposed model. Maruf et al. (2024) focused on classifying freshwater fish diseases in Bangladesh using deep learning techniques. The authors proposed two ensemble models: a baseline Averaged Ensemble (AE) model and a novel Performance Metric-Infused Weighted Ensemble (PMIWE) model. The study achieved a testing accuracy of 97.53%, with precision, recall and F1-score all at 97%. To enhance the inter-pretability and trustworthiness of the model, the authors employed Grad-CAM (Gradient-weighted Class Activation Mapping), an explainable artificial intelligence (XAI) technique. Azhar et al. (2024) focused on detecting Protozoan white spot disease, caused by Cryptocaryon irritans in saltwater fish, using an intelligent system based on a Convolutional Neural Network (CNN). The study used GoogleNet, a deep learning algorithm, to identify infected fish from raw underwater images. The model achieved an accuracy of 90%, offering a promising solution for early screening of this contagious disease.
       
Mia et al. (2022) aimed to improve fish disease recognition to help remote farmers manage their farms. They developed expert systems using smartphone images for disease identification. A segmentation algorithm is selected to distinguish healthy and diseased areas. In this work, eight classification algorithms are tested and achieved a remarkable accuracy of 88.87% with the Random Forest algorithm. Ahmed et al. (2021) conducted a study on fish disease detection using image-based ML techniques in aquaculture. The research had two main phases. The first phase involved image pre-processing and segmentation to reduce noise and enhance image quality. The second phase focused on feature extraction for disease classification using a Support Vector Machine (SVM) algorithm with a kernel function. The processed images were tested on a salmon fish dataset, both with and without image augmentation. The SVM model achieved accuracies of 91.42% with augmentation and 94.12% without augmentation, demonstrating its effectiveness in fish disease detection.
               
In summary, the VGGNet19 model demonstrates high accuracy in fish disease detection, with promising classification performance for most disease categories. However, challenges remain in classifying visually similar diseases, particularly White spot disease and Saprolegniasis. Addressing these limitations through advanced model optimization techniques could further enhance disease detection capabilities and improve practical applicability in aquaculture health monitoring systems.
This study proposed a fish disease detection approach using VGGNet-19, achieving 90.96% accuracy across seven disease categories. While the model performed well, it struggled with White Spot and Saprolegniasis, requiring further fine-tuning. The dataset provides a structured representation of specific diseases but does not cover all fish diseases. The flexible VGGNet-19 architecture allows for expansion with more labeled data. Future work will enhance accuracy by exploring ResNet and Inception architectures, using synthetic data and integrating explainable AI. This approach can benefit commercial aquaculture by enabling faster, more accurate disease detection, improving fish health management and reducing economic losses.
This research was funded by the Korea Institute of Marine Science and Technology Promotion (KIMST) funded by the Ministry of Oceans and Fisheries, Korea (RS-2022-KS221676, Development of Digital Flow-through Aquaculture System).
 
Authors’ contributions
 
All authors contributed toward data analysis, drafting and revising the paper and agreed to be responsible for all the aspects of this work.
 
Data availability
 
The data analysed/generated in the present study will be made available from corresponding authors upon reasonable request.
 
Use of artificial intelligence
 
Not applicable.
 
Declarations
 
Authors declare that all works are original and this manuscript has not been published in any other journal.
Authors declare that they have no conflict of interest.

  1. Ahmed, M.S., Aurpa, T.T. and Azad, M.A.K. (2021). Fish disease detection using image-based machine learning technique in aquaculture. Journal of King Saud University - Computer and Information Sciences. 34(8): 5170-5182. https://doi. org/10.1016/j.jksuci.2021.05.003.

  2. AlZubi, A.A. (2023). Artificial Intelligence and its Application in the prediction and diagnosis of animal diseases: A review. Indian Journal of Animal Research. 57(10): 1265-1271. doi: 10.18805/IJAR.BF-1684.

  3. Azhar, N.A.S.B.M., Harun, N.N.H.B., Hassan, N.M.G.B., Yusoff, N.N.B., Pauzi, N.S.N.B.M., Yusuf, N.N.N. and Chu, N.K.B. (2024). Early screening protozoan white spot fish disease using convolutional neural network. Journal of Advanced Research in Applied Sciences and Engineering Technology. 37(1): 49-55. https://doi.org/10.37934/araset.37.1. 4955.

  4. Bagga, T., Ansari, A.H., Akhter, S., Mittal, A. and Mittal, A. (2024). Understanding Indian consumers’ propensity to purchase Electric vehicles: An Analysis of Determining Factors in Environmentally Sustainable Transportation. International Journal of Environmental Sciences. 10(1): 1-13. https:// www.theaspd.com/resources/1.%20Electric% 20Vehicles% 20and%20Enviorment.pdf.

  5. Begum, R., Himaya, S.M.M.S., Imthiyas, M.S.M. and Afreen, S.M.M.S. (2024). Fish waste: Understanding the Pollution Potential and Sustainable Mitigation Strategies. In S. Maqsood, M. N. Naseer, S. Benjakul and A. A. Zaidi (Eds.), Fish Waste to Valuable Products: Sustainable Materials and Technology. (pp. 401-420). Springer. https://doi.org/10.1007/978-981- 99-8593-7_20.

  6. Cho, O.H. (2024). An evaluation of various machine learning approaches for detecting leaf diseases in Agriculture. Legume Research. 47(4): 619-627. https://doi.org/10. 18805/LRF-787

  7. Coppola, D., Lauritano, C., Palma Esposito, F., Riccio, G., Rizzo, C. and de Pascale, D. (2021). Fish waste: From problem to valuable resource. Marine Drugs. 19(2): 116. https:// doi.org/10.3390/md19020116.

  8. Dataset Source: https://www.kaggle.com/datasets/subirbiswas19/ freshwater-fish-disease-aquaculture-in-south-asia? select=SB-FishDisease

  9. FAO. (2020). The State of World Fisheries and Aquaculture. Available online: https://www.fao.org/interactive/state-of-fisheries- aquaculture/2020/en/.

  10. FAO. (2023). Global fisheries and aquaculture production reaches a new record high. Retrieved from https://www.fao.org/ newsroom/detail/fao-report-global-fisheries-and- aquaculture-production-reaches-a-new-record-high/ en.

  11. Hai, N.T. and Duong, N.T. (2024). An improved environmental management model for assuring energy and economic prosperity. Acta Innovations. 52: 9-18. https://doi.org/ 10.62441/ActaInnovations.52.2. 

  12. Isaac, S., Daniel, O., Grivin, M., Kyagaba, J.M., Lule, E. and Marvin, G. (2024). Explainable machine vision techniques for fish disease detection with deep transfer learning. 2024 5th International Conference on Electronics and Sustainable Communication Systems (ICESC). Coimbatore, India. 1218-1227. https://doi.org/10.1109/ICESC60852.2024. 10689742.

  13. Islam, S.I., Ahammad, F. and Mohammed, H. (2024). Cutting-edge technologies for detecting and controlling fish diseases: Current status, outlook and challenges. Journal of the World Aquaculture Society. 55(2): e13051. https://doi.org/ 10.1111/jwas.13051.

  14. Jiang, Z., Wu, H., Li, X., Ma, C. and Liu, C. (2024). Research on fish disease detection based on big data and deep learning. BDSIC ’24: Proceedings of the 2024 6th International Conference on Big-data Service and Intelligent Computation. 57-61. https://doi.org/10.1145/3686540.3686548.

  15. Maltare, N.N., Sharma, D. and Patel, S. (2023). An exploration and prediction of rainfall and groundwater level for the district of Banaskantha, Gujrat, India. International Journal of Environmental Sciences. 9(1): 1-17. https://www.theaspd. com/resources/v9-1-1-Nilesh%20N.%20Maltare.pdf.

  16. Maruf, A.A., Fahim, S.H., Bashar, R., Rumy, R.A., Chowdhury, S.I. and Aung, Z. (2024). Classification of freshwater fish diseases in Bangladesh using a novel ensemble deep learning model: Enhancing accuracy and interpretability. IEEE Access. 12: 96411-96435. https://doi.org/10.1109/ACCESS. 2024.3426041.

  17. Mia, M.J., Mahmud, R.B., Sadad, M.S., Asad, H.A. and Hossain, R. (2022). An in-depth automated approach for fish disease recognition. Journal of King Saud University - Computer and Information Sciences. 34(9): 7174-7183. https:// doi.org/10.1016/j.jksuci.2022.02.023.

  18. Nguyen, H.V., Huynh, T.Q., Nguyen, N.M., Su, A.K. and Nguyen, H.T. (2024). Comparative Analysis of Fine-tuned MobileNet Versions on Fish Disease Detection. In L. Barolli (Ed.), Innovative Mobile and Internet Services in Ubiquitous Computing. Lecture Notes on Data Engineering and Communications Technologies. 214. (pp. 1-12). Springer. https://doi.org/10.1007/978-3-031-64766-6_20.

  19. Statista. (2024). Total world fish production from 2002 to 2023. Retrieved from https://www.statista.com/statistics/ 264577/total-world-fish-production-since-2002/.

  20. Sven, L., Timmerhaus, G., Johansen, L. and Ytteborg, E. (2021). Deep neural network analysis - a paradigm shift for histological examination of health and welfare of farmed fish. Aqua- Culture. 532: 736024. https://doi.org/10.1016/j.aquaculture. 2020.736024.

  21. Vasumathi, A., Rathore, P.S., Bharathi, S.E., N, V. and S.H. (2024). Fish disease detection using machine learning. 2024 International Conference on Science Technology Engineering and Management (ICSTEM), Coimbatore, India. 1-4. https://doi.org/10.1109/ICSTEM61137.2024.10560522.

  22. World Aquaculture Society. (2023). Aquaculture production statistics. Retrieved from https://www.was.org/Magazine/Fetch Article/1895.

  23. Zhang, X., Zhang, W., Zhang, C., Zhang, L. and Zhang, Y. (2024). A multi-indexes and non-invasive fish health assessment system with deep learning and impedance sensing. Aquaculture. 598: 742025. https://doi.org/10.1016/ j.aqua culture.2024.742025.

Editorial Board

View all (0)