Deep Learning for Disease Detection in Cauliflower Plants

In an article published in the journal Nature, the authors used Convolutional Neural Networks (CNN) to identify and detect various diseases in cauliflower crops. This research aimed to protect agricultural yields and human health by preventing the consumption of contaminated produce.

Study: Deep Learning for Disease Detection in Cauliflower Plants. Image credit: Generated using DALL.E.3
Study: Deep Learning for Disease Detection in Cauliflower Plants. Image credit: Generated using DALL.E.3

Background

Cauliflower, a vital crop in agriculture, is susceptible to various diseases that not only affect crop yields but also pose health risks to consumers. Detection and classification of these cauliflower diseases have garnered significant attention due to their economic and health implications. Several studies have delved into using advanced technologies to address this issue.

Existing research has explored the use of techniques like K-means clustering, CNNs, and traditional machine learning methods to identify and classify cauliflower diseases. These studies leveraged datasets, including VegNet, to develop algorithms for disease detection. Notably, some studies achieved high accuracy rates, with models like InceptionV3 demonstrating remarkable performance in disease classification.

Many studies used relatively small datasets, leading to concerns about class imbalances and the need for further fine-tuning to enhance accuracy. Additionally, the computational demands of certain models and the lack of diverse datasets hinder their practical implementation, and there are concerns about the limited generalization of these models.

The authors of the present study aimed to address these gaps by leveraging advanced deep-learning models for cauliflower disease detection. By combining these models, the study endeavored to create a more robust and widely applicable cauliflower disease detection system. The study also recognized the importance of reducing data dependency and sought to develop a model that can operate in real time and overcome the computational limitations of previous approaches. Overall, this research strived to enhance cauliflower production efficiency and safeguard both crop yields and human health.

Materials and Methods

Dataset: For this research, two sets of cauliflower images were used: an original image set and an augmented image set encompassing three disease classes and disease-free cauliflower. These images were collected from the Manikganj region in Bangladesh. In total, the dataset comprised 66 original images and 7,360 augmented images, categorized into various classes.

Data Pre-processing: The images, initially of size 224x224, underwent pre-processing. Using OpenCV, they were resized while preserving the aspect ratio. Grayscale conversion was applied to simplify the image structure. Morphological operations such as dilation and erosion were performed to manipulate the boundary pixels.

Exploratory Data Analysis: Visual representation of image pixel intensity was carried out. Histograms of red, green, and blue colors were presented to quantitatively show pixel intensities. Histogram Equalization (HE) expanded the intensity range, enhancing image contrast.

Feature Extraction: Several parameters were extracted from the images. These included area, aspect ratio, perimeter, equivalent diameter, max and min values, locations, and mean color intensity. Extreme points and contours were determined, and cropping was performed based on the largest contour. Adaptive thresholding was executed for image segmentation, where different threshold values were calculated for various image regions.

Data Division: The dataset was divided into training and validation sets in a 3:1 ratio. The training set contained 1,384, 1,440, 1,648, and 1,416 images from the bacterial spot rot, black rot, downy mildew, and no disease classes, respectively. The validation set consisted of 346, 360, 412, and 354 images from the same classes.

Applied Models: Several models were applied to detect and classify cauliflower diseases, including Xception, EfficientNet B0, B1, B2, B3, B4, MobileNetV2, DenseNet201, ResNet152V2, and InceptionResNetV2. Various evaluation metrics assessed their performance, including accuracy, loss, root mean square error (RMSE), precision, recall, and F1 score.

Study Findings

The evaluation results showed that EfficientNetB1 achieved the highest accuracy (99.37%) during training, while Xception had the best loss and RMSE values. On the other hand, ResNet152V2 had the lowest accuracy (52.42%) during training.
Additionally, for specific parameters and classes, models exhibited variations in performance. For instance, Xception showed high precision, recall, and F1 score for most classes. EfficientNetB1 performed well in several classes regarding accuracy, precision, recall, and F1 score. DenseNet201 also demonstrated strong performance, especially for recall, precision, and F1 score, while MobileNetV2 had relatively lower performance, particularly for black rot and no disease classes.

The confusion matrix and corresponding metrics were computed for different disease classes (bacterial spot rot, black rot, downy mildew, and no disease), indicating how well the models performed for each class. Overall, EfficientNetB1 and Xception consistently showed strong performance in various classes. The computational time for model training and evaluation varied, with ResNet152V2 taking the longest (1 hour and 36 minutes) and EfficientNetB0 the shortest (24 minutes and 49 seconds).
Comparing the accuracy of the proposed system with existing techniques, the research achieved higher accuracy (99.90% with EfficientNetB1) compared to previous methods. EfficientNetB1 outperformed all other models in the study.

Conclusion

In conclusion, this research illustrates how technology can enhance agriculture by developing an efficient and accurate cauliflower disease detection system. It utilized 7,360 images from four classes, applying image preprocessing and contour feature extraction techniques. Ten models were trained, with EfficientNetB1 achieving the highest validation accuracy of 99.90%.

While some limitations were noted, including potential misclassification, future work will focus on improving contour feature extraction and expanding the system to detect diseases in various vegetables. Overall, this research shows how deep learning models can boost agriculture, increase yields, and ensure global food security. The ongoing refinements promise a lasting positive impact on farming practices.

Journal reference:
Soham Nandi

Written by

Soham Nandi

Soham Nandi is a technical writer based in Memari, India. His academic background is in Computer Science Engineering, specializing in Artificial Intelligence and Machine learning. He has extensive experience in Data Analytics, Machine Learning, and Python. He has worked on group projects that required the implementation of Computer Vision, Image Classification, and App Development.

Citations

Please use one of the following formats to cite this article in your essay, paper or report:

  • APA

    Nandi, Soham. (2023, October 31). Deep Learning for Disease Detection in Cauliflower Plants. AZoAi. Retrieved on November 21, 2024 from https://www.azoai.com/news/20231031/Deep-Learning-for-Disease-Detection-in-Cauliflower-Plants.aspx.

  • MLA

    Nandi, Soham. "Deep Learning for Disease Detection in Cauliflower Plants". AZoAi. 21 November 2024. <https://www.azoai.com/news/20231031/Deep-Learning-for-Disease-Detection-in-Cauliflower-Plants.aspx>.

  • Chicago

    Nandi, Soham. "Deep Learning for Disease Detection in Cauliflower Plants". AZoAi. https://www.azoai.com/news/20231031/Deep-Learning-for-Disease-Detection-in-Cauliflower-Plants.aspx. (accessed November 21, 2024).

  • Harvard

    Nandi, Soham. 2023. Deep Learning for Disease Detection in Cauliflower Plants. AZoAi, viewed 21 November 2024, https://www.azoai.com/news/20231031/Deep-Learning-for-Disease-Detection-in-Cauliflower-Plants.aspx.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of AZoAi.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Revolutionizing Gemstone Analysis with Deep Learning