Arşiv logosu
  • Türkçe
  • English
  • Giriş
    Yeni kullanıcı mısınız? Kayıt için tıklayın. Şifrenizi mi unuttunuz?
Arşiv logosu
  • Koleksiyonlar
  • Sistem İçeriği
  • Analiz
  • Talep/Soru
  • Türkçe
  • English
  • Giriş
    Yeni kullanıcı mısınız? Kayıt için tıklayın. Şifrenizi mi unuttunuz?
  1. Ana Sayfa
  2. Yazara Göre Listele

Yazar "Mowla, Neazmul" seçeneğine göre listele

Listeleniyor 1 - 2 / 2
Sayfa Başına Sonuç
Sıralama seçenekleri
  • [ X ]
    Öğe
    A Lightweight Deep Learning Model for Retinopathy of Prematurity Classification in eHealth Applications
    (IEEE, 2025) Mowla, Neazmul; Mowla, Md. Najmul; Rabie, Khaled; Alsinglawi, Belal
    Retinopathy of Prematurity (ROP) is a vision-threatening condition in premature infants requiring timely and accurate diagnosis to prevent blindness. While electronic health (eHealth) technologies promise to improve neonatal care, automating ROP diagnosis faces challenges such as limited labeled datasets, architectural complexity, and high computational demands. This study introduces LightEyeNet, a lightweight deep-learning architecture optimized for eHealth applications in ROP severity classification. By integrating DenseNet121 and a channel-wise residual attention network block, LightEyeNet enhances diagnostic accuracy and efficiency. Explainable AI techniques, including Grad-CAM and LIME, further improve transparency and clinical interpretability. LightEyeNet achieves 96.28% testing accuracy, outperforming state-of-the-art pre-trained networks, including DenseNet201 (95.78%, + 0.5%), Inception-V3 (93.80%, + 2.48%), Xception (94.54%, + 1.74%), and EfficientDense (87.10%, + 9.18%). Furthermore, LightEyeNet is the most compact architecture among these, with a size of 2.45 MB, compared to Efficient-Dense (6.88 MB), Inception-V3 (3.56 MB), Xception (21.48 MB), and DenseNet201 (5.05 MB). With a specificity of 0.99, a sensitivity of 0.95, and an AUC score of 0.99 across five ROP severity classes, LightEyeNet demonstrates a balance of superior performance and efficiency.
  • [ X ]
    Öğe
    Internet of Things and Wireless Sensor Networks for Smart Agriculture Applications: A Survey
    (IEEE-Inst Electrical Electronics Engineers Inc, 2023) Mowla, Md. Najmul; Mowla, Neazmul; Shah, A. F. M. Shahen; Rabie, Khaled M.; Shongwe, Thokozani
    The increasing food scarcity necessitates sustainable agriculture achieved through automation to meet the growing demand. Integrating the Internet of Things (IoT) and Wireless Sensor Networks (WSNs) is crucial in enhancing food production across various agricultural domains, encompassing irrigation, soil moisture monitoring, fertilizer optimization and control, early-stage pest and crop disease management, and energy conservation. Wireless application protocols such as ZigBee, WiFi, SigFox, and LoRaWAN are commonly employed to collect real-time data for monitoring purposes. Embracing advanced technology is imperative to ensure efficient annual production. Therefore, this study emphasizes a comprehensive, future-oriented approach, delving into IoT-WSNs, wireless network protocols, and their applications in agriculture since 2019. It thoroughly discusses the overview of IoT and WSNs, encompassing their architectures and summarization of network protocols. Furthermore, the study addresses recent issues and challenges related to IoT-WSNs and proposes mitigation strategies. It provides clear recommendations for the future, emphasizing the integration of advanced technology aiming to contribute to the future development of smart agriculture systems.

| Adana Alparslan Türkeş Bilim ve Teknoloji Üniversitesi | Kütüphane | Rehber | OAI-PMH |

Bu site Creative Commons Alıntı-Gayri Ticari-Türetilemez 4.0 Uluslararası Lisansı ile korunmaktadır.


Balcalı Mahallesi, Güney Kampüs, 10. Sokak, No: 1U, Sarıçam, Adana, TÜRKİYE
İçerikte herhangi bir hata görürseniz lütfen bize bildirin

DSpace 7.6.1, Powered by İdeal DSpace

DSpace yazılımı telif hakkı © 2002-2026 LYRASIS

  • Çerez Ayarları
  • Gizlilik Politikası
  • Son Kullanıcı Sözleşmesi
  • Geri Bildirim