TY - JOUR
T1 - Artificial Intelligence-Based Plant Disease Classification in Low-Light Environments
AU - Gondal, Hafiz Ali Hamza
AU - Jeong, Seong In
AU - Jang, Won Ho
AU - Kim, Jun Seo
AU - Akram, Rehan
AU - Irfan, Muhammad
AU - Tariq, Muhammad Hamza
AU - Park, Kang Ryoung
N1 - Publisher Copyright:
© 2025 by the authors.
PY - 2025/11
Y1 - 2025/11
N2 - The accurate classification of plant diseases is vital for global food security, as diseases can cause major yield losses and threaten sustainable and precision agriculture. The classification of plant diseases in low-light noisy environments is crucial because crops can be continuously monitored even at night. Important visual cues of disease symptoms can be lost due to the degraded quality of images captured under low-illumination, resulting in poor performance of conventional plant disease classifiers. However, researchers have proposed various techniques for classifying plant diseases in daylight, and no studies have been conducted for low-light noisy environments. Therefore, we propose a novel model for classifying plant diseases from low-light noisy images called dilated pixel attention network (DPA-Net). DPA-Net uses a pixel attention mechanism and multi-layer dilated convolution with a high receptive field, which obtains essential features while highlighting the most relevant information under this challenging condition, allowing more accurate classification results. Additionally, we performed fractal dimension estimation on diseased and healthy leaves to analyze the structural irregularities and complexities. For the performance evaluation, experiments were conducted on two public datasets: the PlantVillage and Potato Leaf Disease datasets. In both datasets, the image resolution is 256 × 256 pixels in joint photographic experts group (JPG) format. For the first dataset, DPA-Net achieved an average accuracy of 92.11% and harmonic mean of precision and recall (F1-score) of 89.11%. For the second dataset, it achieved an average accuracy of 88.92% and an F1-score of 88.60%. These results revealed that the proposed method outperforms state-of-the-art methods. On the first dataset, our method achieved an improvement of 2.27% in average accuracy and 2.86% in F1-score compared to the baseline. Similarly, on the second dataset, it attained an improvement of 6.32% in average accuracy and 6.37% in F1-score over the baseline. In addition, we confirm that our method is effective with the real low-illumination dataset self-constructed by capturing images at 0 lux using a smartphone at night. This approach provides farmers with an affordable practical tool for early disease detection, which can support crop protection worldwide.
AB - The accurate classification of plant diseases is vital for global food security, as diseases can cause major yield losses and threaten sustainable and precision agriculture. The classification of plant diseases in low-light noisy environments is crucial because crops can be continuously monitored even at night. Important visual cues of disease symptoms can be lost due to the degraded quality of images captured under low-illumination, resulting in poor performance of conventional plant disease classifiers. However, researchers have proposed various techniques for classifying plant diseases in daylight, and no studies have been conducted for low-light noisy environments. Therefore, we propose a novel model for classifying plant diseases from low-light noisy images called dilated pixel attention network (DPA-Net). DPA-Net uses a pixel attention mechanism and multi-layer dilated convolution with a high receptive field, which obtains essential features while highlighting the most relevant information under this challenging condition, allowing more accurate classification results. Additionally, we performed fractal dimension estimation on diseased and healthy leaves to analyze the structural irregularities and complexities. For the performance evaluation, experiments were conducted on two public datasets: the PlantVillage and Potato Leaf Disease datasets. In both datasets, the image resolution is 256 × 256 pixels in joint photographic experts group (JPG) format. For the first dataset, DPA-Net achieved an average accuracy of 92.11% and harmonic mean of precision and recall (F1-score) of 89.11%. For the second dataset, it achieved an average accuracy of 88.92% and an F1-score of 88.60%. These results revealed that the proposed method outperforms state-of-the-art methods. On the first dataset, our method achieved an improvement of 2.27% in average accuracy and 2.86% in F1-score compared to the baseline. Similarly, on the second dataset, it attained an improvement of 6.32% in average accuracy and 6.37% in F1-score over the baseline. In addition, we confirm that our method is effective with the real low-illumination dataset self-constructed by capturing images at 0 lux using a smartphone at night. This approach provides farmers with an affordable practical tool for early disease detection, which can support crop protection worldwide.
KW - Internet-of-Things
KW - artificial intelligence
KW - crop protection
KW - dilated pixel attention network
KW - fractal dimension estimation
UR - https://www.scopus.com/pages/publications/105023092431
U2 - 10.3390/fractalfract9110691
DO - 10.3390/fractalfract9110691
M3 - Article
AN - SCOPUS:105023092431
SN - 2504-3110
VL - 9
JO - Fractal and Fractional
JF - Fractal and Fractional
IS - 11
M1 - 691
ER -