Repository logo
Communities & Collections
All of DSpace
  • English
  • العربية
  • বাংলা
  • Català
  • Čeština
  • Deutsch
  • Ελληνικά
  • Español
  • Suomi
  • Français
  • Gàidhlig
  • हिंदी
  • Magyar
  • Italiano
  • Қазақ
  • Latviešu
  • Nederlands
  • Polski
  • Português
  • Português do Brasil
  • Srpski (lat)
  • Српски
  • Svenska
  • Türkçe
  • Yкраї́нська
  • Tiếng Việt
Log In
New user? Click here to register.Have you forgotten your password?
  1. Home
  2. Browse by Author

Browsing by Author "Vidhanaarachchi, Samitha P."

Filter results by typing the first few letters
Now showing 1 - 1 of 1
  • Results Per Page
  • Sort Options
  • No Thumbnail Available
    Item
    Effectiveness of Using Deep Learning for Blister Blight Identification in Sri Lankan Tea
    (Department of Industrial Management, Faculty of Science, University of Kelaniya Sri Lanka, 2023) Hewawitharana, G.H.A.U.; Nawarathne, U.M.L.A.; Hassan, A.S.F.; Wijerathna, L.M.; Sinniah, Ganga D; Vidhanaarachchi, Samitha P.; Wickramarathne, Jagath; Wijekoon, Janaka L.
    Ceylon tea industry faces a major challenge in the form of pathogen-induced crop loss, with Blister Blight (BB) caused by Exobasidium vexans posing the greatest threat, leading to harvest losses of over 30%. This fungus attacks the tender tea shoots, resulting in a direct negative impact on the tea harvest. This paper presents a system to identify the suspicious tea leaves and BB disease at its early stages along with an assessment of severity, offering a potential solution to this critical issue. By utilizing real-time object detection, the system filters out non-tea leaves from the captured initial image of a segment of a tea plant. The identified tea leaves are then subjected to BB identification and severity assessment based on differing visual symptoms of the BB stages. This approach enables the system to accurately identify BB in the initial stage and severity stage, allowing for timely and targeted intervention to minimize crop losses. The YOLOv8 model has been able to correctly identify 98% of the objects it has detected as relevant (precision), and it has been able to correctly identify 96% of all the relevant objects present in the scene (recall). The Residual Network 50 (Resnet50) convolutional neural network (CNN) model was selected as the final model, achieving an accuracy of 89.90% during the training phase and an accuracy of 88.26% during the testing phase.

DSpace software copyright © 2002-2025 LYRASIS

  • Privacy policy
  • End User Agreement
  • Send Feedback
Repository logo COAR Notify