Regularized lightweight deep learning for semantic image segmentation
| dc.contributor.advisor | Álvarez Meza, Andrés Marino | |
| dc.contributor.author | Iturriago Salas, Lucas Miguel | |
| dc.contributor.cvlac | Iturriago Salas, Lucas Miguel [0002096549] | |
| dc.contributor.googlescholar | Iturriago Salas, Lucas Miguel [LJMF630AAAAJ&hl] | |
| dc.contributor.orcid | Iturriago Salas, Lucas Miguel [0009000310579095] | |
| dc.contributor.researchgroup | Grupo de Control y Procesamiento Digital de Señales | |
| dc.date.accessioned | 2026-02-27T13:52:27Z | |
| dc.date.available | 2026-02-27T13:52:27Z | |
| dc.date.issued | 2025 | |
| dc.description | ilustraciones, graficas, tablas | spa |
| dc.description.abstract | Semantic image segmentation holds immense potential for transformative applications in critical domains such as healthcare and agriculture. However, the practical deployment of deep learning models is often hindered by three fundamental challenges: the high variability of real-world input data, the cost and inconsistency of annotations, and the computational demands of state-of-the-art architectures. While existing methods achieve high accuracy under controlled conditions, they often lack the robustness and efficiency required to overcome these practical hurdles, limiting their generalization to diverse, resource-constrained environments. This thesis proposes a regularized, lightweight deep learning framework designed to maintain high accuracy and robustness in semantic segmentation across diverse datasets, variable imaging conditions, and application domains, while ensuring efficient deployment in resource-constrained environments. The framework is built on a systematic approach that addresses the entire modeling pipeline, from baseline evaluation to final deployment. To achieve this, three main strategies were developed. First, a comprehensive comparative analysis of canonical segmentation architectures on four heterogeneous datasets was conducted to establish robust baselines and identify the limitations of existing models. This revealed that encoder-decoder architectures like U-Net offer superior generalization but struggle with specific challenges like class imbalance and fine-detail preservation. Second, to tackle annotation noise and disagreement, a novel multi-annotator learning framework, AnnotHarmony, was proposed, centered on a new loss function (TGCESSPS). This approach successfully learns from noisy, sparse, and crowdsourced labels by modeling annotator reliability at the pixel level, outperforming traditional aggregation methods in preserving clinically relevant details. Third, the most effective models were optimized and evaluated for their generalization capacity and computational efficiency. This culminated in the successful deployment of lightweight models on edge devices, including a Raspberry Pi for automated agricultural monitoring and a mobile application for real-time clinical support, demonstrating a practical balance between performance and efficiency. In conclusion, this work bridges the gap between theoretical research and practical application by delivering a holistic framework for developing robust, resilient, and efficient semantic segmentation systems. The methodologies presented advance the state of the art by enabling reliable model training with imperfect data and facilitating the deployment of computer vision solutions in real-world, resource-limited settings, thereby increasing their potential impact in critical fields (Texto tomado de la fuente). | eng |
| dc.description.abstract | La segmentación semántica de imágenes posee un inmenso potencial para aplicaciones transformadoras en dominios críticos como la salud y la agricultura. Sin embargo, el despliegue práctico de modelos de aprendizaje profundo se ve a menudo obstaculizado por tres desafíos fundamentales: la alta variabilidad de los datos de entrada en el mundo real, el costo e inconsistencia de las anotaciones y las exigencias computacionales de las arquitecturas de vanguardia. Aunque los métodos existentes logran una alta precisión bajo condiciones controladas, suelen carecer de la robustez y eficiencia necesarias para superar estos obstáculos prácticos, lo que limita su generalización a entornos diversos y con recursos limitados. Esta tesis propone un marco de aprendizaje profundo ligero y regularizado, diseñado para mantener una alta precisión y robustez en la segmentación semántica a través de diversos conjuntos de datos, condiciones de imagen variables y dominios de aplicación, garantizando al mismo tiempo un despliegue eficiente en entornos de recursos limitados. El marco se basa en un enfoque sistemático que aborda todo el flujo de trabajo del modelado, desde la evaluación base hasta el despliegue final. Para lograr esto, se desarrollaron tres estrategias principales: Primero: Se realizó un análisis comparativo exhaustivo de arquitecturas de segmentación canónicas en cuatro conjuntos de datos heterogéneos para establecer líneas base sólidas e identificar las limitaciones de los modelos existentes. Esto reveló que las arquitecturas de codificador-decodificador (como U-Net) ofrecen una generalización superior, pero presentan dificultades ante desafíos específicos como el desequilibrio de clases y la preservación de detalles finos. Segundo: Para abordar el ruido y las discrepancias en las anotaciones, se propuso un nuevo marco de aprendizaje multi-anotador llamado AnnotHarmony, centrado en una nueva función de pérdida (TGCESSPS). Este enfoque logra aprender con éxito de etiquetas ruidosas, dispersas y obtenidas mediante crowdsourcing al modelar la fiabilidad del anotador a nivel de píxel, superando a los métodos de agregación tradicionales en la preservación de detalles clínicamente relevantes. Tercero: Se optimizaron y evaluaron los modelos más efectivos en cuanto a su capacidad de generalización y eficiencia computacional. Esto culminó en el despliegue exitoso de modelos ligeros en dispositivos de borde (edge devices), incluyendo una Raspberry Pi para el monitoreo agrícola automatizado y una aplicación móvil para soporte clínico en tiempo real, demostrando un equilibrio práctico entre rendimiento y eficiencia. En conclusión, este trabajo cierra la brecha entre la investigación teórica y la aplicación práctica al entregar un marco integral para el desarrollo de sistemas de segmentación semántica robustos, resilientes y eficientes. Las metodologías presentadas avanzan el estado del arte al permitir el entrenamiento de modelos confiables con datos imperfectos y facilitar el despliegue de soluciones de visión por computadora en entornos reales de recursos limitados, aumentando así su impacto potencial en campos críticos. | spa |
| dc.description.curriculararea | Eléctrica, Electrónica, Automatización Y Telecomunicaciones.Sede Manizales | |
| dc.description.degreelevel | Maestría | |
| dc.description.degreename | Magíster en Ingeniería - Automatización Industrial | |
| dc.description.researcharea | Inteligencia artificial | |
| dc.format.extent | xx, 137 páginas | |
| dc.format.mimetype | application/pdf | |
| dc.identifier.instname | Universidad Nacional de Colombia | spa |
| dc.identifier.reponame | Repositorio Institucional Universidad Nacional de Colombia | spa |
| dc.identifier.repourl | https://repositorio.unal.edu.co/ | spa |
| dc.identifier.uri | https://repositorio.unal.edu.co/handle/unal/89697 | |
| dc.language.iso | eng | |
| dc.publisher | Universidad Nacional de Colombia | |
| dc.publisher.branch | Universidad Nacional de Colombia - Sede Manizales | |
| dc.publisher.faculty | Facultad de Ingeniería y Arquitectura | |
| dc.publisher.place | Manizales, Colombia | |
| dc.publisher.program | Manizales - Ingeniería y Arquitectura - Maestría en Ingeniería - Automatización Industrial | |
| dc.relation.indexed | Agrosavia | |
| dc.relation.indexed | Bireme | |
| dc.relation.indexed | RedCol | |
| dc.relation.indexed | LaReferencia | |
| dc.relation.indexed | Agrovoc | |
| dc.relation.references | Appen, “Computer vision vs. machine vision comparison [image].” https://www.appen.com/blog/computer-vision-vs-machine-vision, 2019. Accessed August 5, 2025. Image published in Appen Blog: "Computer Vision vs. Machine Vision — What’s the Difference?" | |
| dc.relation.references | Market.us Scoop, “Computer vision market size share by industry,” 2024. Accessed: 2025-08-05 | |
| dc.relation.references | Global Market Insights, “Ai in computer vision market size, by end use, 2022–2034,” 2024. Accessed: 2025-08-05 | |
| dc.relation.references | Y. Liu, P. Ge, Q. Liu, S. Fan, and Y. Wang, “An empirical study on multi-domain robust semantic segmentation,” International Journal of Computer Vision, vol. 132, no. 10, pp. 4289–4304, 2024 | |
| dc.relation.references | Hopstarter. https://www.flaticon.com/authors/hopstarter, 2025. Free icons licensed under Flaticon usage terms; attribution required for Free users | |
| dc.relation.references | HAJICON. https://www.flaticon.com/authors/hajicon, 2025. Free icons licensed under Flaticon usage terms; attribution required for Free users | |
| dc.relation.references | S. Kazemifar, A. Balagopal, D. Nguyen, S. McGuire, R. Hannan, S. Jiang, and A. Owrangi, “Segmentation of the prostate and organs at risk in male pelvic ct images using deep learning,” Biomedical Physics & Engineering Express, vol. 4, no. 5, p. 055003, 2018 | |
| dc.relation.references | TagX, “Computer vision and data annotation leading the way for drone ai.” https://medium.com/@tagx20/computer-vision-and-data-annotation-leading-the-way-for-drone-ai-ef9070b8fe21, 2023. Accessed: 2025-08-06 | |
| dc.relation.references | I. Sanida, “Crowdsourcing label aggregation: Modeling task and worker correlation.” https://medium.com/sogetiblogsnl/crowdsourcing-label-aggregation-modeling-task-and-worker-correlation-e6ffddc8ae20, 2020. Accessed: 2025-08-06. Image retrieved from article, used for illustrative purposes | |
| dc.relation.references | P. Chlap, H. Min, N. Vandenberg, J. Dowling, L. Holloway, and A. Haworth, “A review of medical image data augmentation techniques for deep learning applications,” Journal of medical imaging and radiation oncology, vol. 65, no. 5, pp. 545–563, 2021 | |
| dc.relation.references | LeewayHertz, “Vision transformer model: Architecture, development and applications.” https://www.leewayhertz.com/vision-transformer-model/, 2025. Accessed: 2025-08-06. Image retrieved from article, used for illustrative purposes | |
| dc.relation.references | N. Sharma, “What is mobilenetv2? features, architecture, application and more.” https://www.analyticsvidhya.com/blog/2023/12/what-is-mobilenetv2/, 2025. Accessed: 2025-08-06. Image retrieved from article, used for illustrative purposes | |
| dc.relation.references | Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” nature, vol. 521, no. 7553, pp. 436–444, 2015 | |
| dc.relation.references | N. Haefner, J. Wincent, V. Parida, and O. Gassmann, “Artificial intelligence and innovation management: A review, framework, and research agenda,” Technological Forecasting and Social Change, vol. 162, p. 120392, 2021 | |
| dc.relation.references | N. Ganesh, R. Shankar, M. Mahdal, J. S. Murugan, J. S. Chohan, and K. Kalita, “Exploring deep learning methods for computer vision applications across multiple sectors: Challenges and future trends.,” CMES-Computer Modeling in Engineering & Sciences, vol. 139, no. 1, 2024 | |
| dc.relation.references | W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and A. C. Berg, “Ssd: Single shot multibox detector,” in European conference on computer vision, pp. 21–37, Springer, 2016 | |
| dc.relation.references | O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, et al., “Imagenet large scale visual recognition challenge,” International journal of computer vision, vol. 115, no. 3, pp. 211–252, 2015 | |
| dc.relation.references | G. Vial, “Understanding digital transformation: A review and a research agenda,” Managing digital transformation, pp. 13–66, 2021 | |
| dc.relation.references | C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1–9, 2015 | |
| dc.relation.references | A. Esteva, B. Kuprel, R. A. Novoa, J. Ko, S. M. Swetter, H. M. Blau, and S. Thrun, “Dermatologist-level classification of skin cancer with deep neural networks,” nature, vol. 542, no. 7639, pp. 115–118, 2017 | |
| dc.relation.references | R. Malhan and S. K. Gupta, “The role of deep learning in manufacturing applications: Challenges and opportunities,” Journal of Computing and Information Science in Engineering, vol. 23, no. 6, p. 060816, 2023 | |
| dc.relation.references | K. B. Singh and M. A. Arat, “Deep learning in the automotive industry: Recent advances and application examples,” arXiv preprint arXiv:1906.08834, 2019 | |
| dc.relation.references | I. Goodfellow, Y. Bengio, A. Courville, and Y. Bengio, Deep learning, vol. 1. MIT press Cambridge, 2016 | |
| dc.relation.references | A. A. Cruz-Roa, J. E. Arevalo Ovalle, A. Madabhushi, and F. A. González Osorio, “A deep learning architecture for image representation, visual interpretability and automated basal-cell carcinoma cancer detection,” in International conference on medical image computing and computer-assisted intervention, pp. 403–410, Springer, 2013 | |
| dc.relation.references | S. A. Harmon, T. H. Sanford, S. Xu, E. B. Turkbey, H. Roth, Z. Xu, D. Yang, A. Myronenko, V. Anderson, A. Amalou, et al., “Artificial intelligence for the detection of covid-19 pneumonia on chest ct using multinational datasets,” Nature communications, vol. 11, no. 1, p. 4080, 2020 | |
| dc.relation.references | A. Rajkomar, J. Dean, and I. Kohane, “Machine learning in medicine,” New England Journal of Medicine, vol. 380, no. 14, pp. 1347–1358, 2019 | |
| dc.relation.references | MarketsandMarkets Research, “Computer vision in healthcare market insights: Size, share, growth, industry trends,” 2024. Accessed: 2025-08-06 | |
| dc.relation.references | Future Market Insights, “Computer vision in healthcare market size, trends & growth 2024–2034.” https://www.futuremarketinsights.com/reports/computer-vision-in-healthcare-market, 2024. Accessed August 2025 | |
| dc.relation.references | H. Lindroth, K. Nalaie, R. Raghu, I. N. Ayala, C. Busch, A. Bhattacharyya, P. Moreno Franco, D. A. Diedrich, B. W. Pickering, and V. Herasevich, “Applied artificial intelligence in healthcare: a review of computer vision technology application in hospital settings,” Journal of Imaging, vol. 10, no. 4, p. 81, 2024 | |
| dc.relation.references | M. I. Madrigal-Garcia, D. Archer, M. Singer, M. Rodrigues, A. Shenfield, and J. Moreno-Cuesta, “Do temporal changes in facial expressions help identify patients at risk of deterioration in hospital wards? a post hoc analysis of the visual early warning score study,” Critical Care Explorations, vol. 2, no. 5, p. e0115, 2020 | |
| dc.relation.references | S. Yeung, F. Rinaldo, J. Jopling, B. Liu, R. Mehra, N. L. Downing, M. Guo, G. M. Bianconi, A. Alahi, J. Lee, et al., “A computer vision system for deep learning-based detection of patient mobilization activities in the icu,” NPJ digital medicine, vol. 2, no. 1, p. 11, 2019 | |
| dc.relation.references | T. Maruyama, N. Hayashi, Y. Sato, S. Hyuga, Y. Wakayama, H. Watanabe, A. Ogura, and T. Ogura, “Comparison of medical image classification accuracy among three machine learning methods,” Journal of X-ray Science and Technology, vol. 26, no. 6, pp. 885–893, 2018 | |
| dc.relation.references | Market Research Future, “Computer vision in healthcare market to project lucrative cagr of 47.3% from 2020-2027,” June 2021. Accessed September 14, 2025 | |
| dc.relation.references | H. Xu, Q. Xu, F. Cong, J. Kang, C. Han, Z. Liu, A. Madabhushi, and C. Lu, “Vision transformers for computational histopathology,” IEEE Reviews in Biomedical Engineering, vol. 17, pp. 63–79, 2023 | |
| dc.relation.references | M. Khened, A. Kori, H. Rajkumar, G. Krishnamurthi, and B. Srinivasan, “A generalized deep learning framework for whole-slide image segmentation and analysis,” Scientific reports, vol. 11, no. 1, p. 11579, 2021 | |
| dc.relation.references | C. Kang, C. Lee, H. Song, M. Ma, and S. Pereira, “Variability matters: Evaluating inter-rater variability in histopathology for robust cell detection,” in European Conference on Computer Vision, pp. 552–565, Springer, 2022 | |
| dc.relation.references | A. A. Munia, M. Abdar, M. Hasan, M. S. Jalali, B. Banerjee, A. Khosravi, I. Hossain, H. Fu, and A. F. Frangi, “Attention-guided hierarchical fusion u-net for uncertainty-driven medical image segmentation,” Information Fusion, vol. 115, p. 102719, 2025 | |
| dc.relation.references | M. Zubair, M. Owais, T. Hassan, M. Bendechache, M. Hussain, I. Hussain, and N. Werghi, “An interpretable framework for gastric cancer classification using multi-channel attention mechanisms and transfer learning approach on histopathology images,” Scientific Reports, vol. 15, no. 1, p. 13087, 2025 | |
| dc.relation.references | E. Mavridou, E. Vrochidou, G. A. Papakostas, T. Pachidis, and V. G. Kaburlasos, “Machine vision systems in precision agriculture for crop farming,” Journal of Imaging, vol. 5, no. 12, p. 89, 2019 | |
| dc.relation.references | Market.US, “Global ai in agriculture market by technology, application, component, and region– industry segment outlook, market assessment, competition scenario, trends, and forecast 2023–2032.” Online Report, April 2024. Accessed September 14, 2025 | |
| dc.relation.references | GlobalMarketInsights Inc., “Precision farming market size by component, technology, application, and farm size, growth forecast 2025–2034.” Online Market Report, January 2025. Accessed September 14, 2025 | |
| dc.relation.references | A. Yeshe, P. Gourkhede, and P. Vaidya, “Blue river technology: Futuristic approach of precision farming,” Just Agriculture: Punjab, India, 2022 | |
| dc.relation.references | Grand View Research, “Agriculture drones market size, share & trends analysis report by type (fixed wing, rotary wing), by component (hardware, software), by farming environment (indoor, outdoor), by application, by region, and segment forecasts, 2025- 2030.” Online Market Report, April 2025. Accessed September 14, 2025 | |
| dc.relation.references | J. Colmer, C. M. O’Neill, R. Wells, A. Bostrom, D. Reynolds, D. Websdale, G. Shiralagi, W. Lu, Q. Lou, T. Le Cornu, et al., “Seedgerm: a cost-effective phenotyping platform for automated seed imaging and machine-learning based phenotypic analysis of crop seed germination,” New Phytologist, vol. 228, no. 2, pp. 778–793, 2020 | |
| dc.relation.references | N. Genze, R. Bharti, M. Grieb, S. J. Schultheiss, and D. G. Grimm, “Accurate machine learning-based germination detection, prediction and quality assessment of three grain crops,” Plant methods, vol. 16, no. 1, p. 157, 2020 | |
| dc.relation.references | S. Samiei, P. Rasti, J. Ly Vu, J. Buitink, and D. Rousseau, “Deep learning-based detection of seedling development,” Plant Methods, vol. 16, no. 1, p. 103, 2020 | |
| dc.relation.references | Q. Peng, L. Tu, Y. Wu, Z. Yu, G. Tang, and W. Song, “Automatic monitoring system for seed germination test based on deep learning,” Journal of Electrical and Computer Engineering, vol. 2022, no. 1, p. 4678316, 2022 | |
| dc.relation.references | A. Voulodimos, N. Doulamis, A. Doulamis, and E. Protopapadakis, “Deep learning for computer vision: A brief review,” Computational intelligence and neuroscience, vol. 2018, no. 1, p. 7068349, 2018 | |
| dc.relation.references | C. Shorten and T. M. Khoshgoftaar, “A survey on image data augmentation for deep learning,” Journal of Big Data, vol. 6, no. 1, pp. 1–48, 2019 | |
| dc.relation.references | A. Alruwaili and M. Alsalim, “Data diversity and its impact on machine learning fairness,” International Journal of Cloud Computing and Database Management, vol. 5, no. 1, pp. 38–41, 2024 | |
| dc.relation.references | H. Javed, S. El-Sappagh, and T. Abuhmed, “Robustness in deep learning models for medical diagnostics: security and adversarial challenges towards robust ai applications,” Artificial Intelligence Review, vol. 58, no. 1, p. 12, 2024 | |
| dc.relation.references | L. Alzubaidi, J. Bai, A. Al-Sabaawi, J. Santamaría, A. Albahri, B. Al-dabbagh, M. Fadhel, M. Manoufali, J. Zhang, A. Al-Timemy, et al., “A survey on deep learning tools dealing with data scarcity: definitions, challenges, solutions, tips, and applications. j big data 10 (1): 46,” 2023 | |
| dc.relation.references | X. Feng, Y. Jiang, X. Yang, M. Du, and X. Li, “Computer vision algorithms and hardware implementations: A survey,” Integration, vol. 69, pp. 309–320, 2019 | |
| dc.relation.references | Z. Wan, Z. Wang, C. Chung, and Z. Wang, “A survey of dataset refinement for problems in computer vision datasets,” arXiv preprint arXiv:2210.11717, 2023 | |
| dc.relation.references | J. Skibicki, A. Golijanek-Jędrzejczyk, and A. Dzwonkowski, “The influence of camera and optical system parameters on the uncertainty of object location measurement in vision systems,” Sensors, vol. 20, no. 18, p. 5433, 2020 | |
| dc.relation.references | K. Liao, L. Nie, S. Huang, C. Lin, J. Zhang, Y. Zhao, M. Gabbouj, and D. Tao, “Deep learning for camera calibration and beyond: A survey,” arXiv preprint arXiv:2303.10559, 2023 | |
| dc.relation.references | E. C. Covert, K. Fitzpatrick, J. Mikell, R. K. Kaza, J. D. Millet, D. Barkmeier, J. Gemmete, J. Christensen, M. J. Schipper, and Y. K. Dewaraja, “Intra-and inter-operator variability in mri-based manual segmentation of hcc lesions and its impact on dosimetry,” EJNMMI physics, vol. 9, no. 1, p. 90, 2022 | |
| dc.relation.references | F. Renard, S. Guedria, N. De Palma, and N. Vuillerme, “Variability and reproducibility in deep learning for medical image segmentation,” Scientific Reports, vol. 10, no. 1, p. 13724, 2020 | |
| dc.relation.references | M. Moayeri, P. Pope, Y. Balaji, and S. Feizi, “A comprehensive study of image classification model sensitivity to foregrounds, backgrounds, and visual attributes,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 19087–19097, 2022 | |
| dc.relation.references | M. Li, Y. Jiang, Y. Zhang, and H. Zhu, “Medical image analysis using deep learning algorithms,” Frontiers in Public Health, vol. 11, p. 1273253, 2023 | |
| dc.relation.references | N. Sourlos, R. Vliegenthart, J. Santinha, M. E. Klontzas, R. Cuocolo, M. Huisman, and P. van Ooijen, “Recommendations for the creation of benchmark datasets for reproducible artificial intelligence in radiology,” Insights into Imaging, vol. 15, no. 1, p. 248, 2024 | |
| dc.relation.references | S. Hao, W. Han, T. Jiang, Y. Li, H. Wu, C. Zhong, Z. Zhou, and H. Tang, “Synthetic data in ai: Challenges, applications, and ethical implications,” arXiv preprint arXiv:2401.01629, 2024 | |
| dc.relation.references | K. Man and J. Chahl, “A review of synthetic image data and its use in computer vision,” J Imaging, vol. 8, no. 11, p. 310, 2022 | |
| dc.relation.references | S. Yang, W. Xiao, M. Zhang, S. Guo, J. Zhao, and F. Shen, “Image data augmentation for deep learning: A survey,” arXiv preprint arXiv:2204.08610, 2022 | |
| dc.relation.references | S. I. Mirzadeh, M. Farajtabar, R. Pascanu, and H. Ghasemzadeh, “Understanding the role of training regimes in continual learning,” Advances in Neural Information Processing Systems, vol. 33, pp. 7308–7320, 2020 | |
| dc.relation.references | R. Schäfer, T. Nicke, H. Höfener, A. Lange, D. Merhof, F. Feuerhake, V. Schulz, J. Lotz, and F. Kiessling, “Overcoming data scarcity in biomedical imaging with a foundational multi-task model,” Nature Computational Science, vol. 4, no. 7, pp. 495–509, 2024 | |
| dc.relation.references | M. J. Willemink, W. A. Koszek, C. Hardell, J. Wu, D. Fleischmann, H. Harvey, L. R. Folio, R. M. Summers, D. L. Rubin, and M. P. Lungren, “Preparing medical imaging data for machine learning,” Radiology, vol. 295, no. 1, pp. 4–15, 2020 | |
| dc.relation.references | D. Joshi and C. Witharana, “Vision transformer-based unhealthy tree crown detection in mixed northeastern us forests and evaluation of annotation uncertainty,” Remote Sensing, vol. 17, no. 6, p. 1066, 2025 | |
| dc.relation.references | C. G. Northcutt, A. Athalye, and J. Mueller, “Pervasive label errors in test sets destabilize machine learning benchmarks,” arXiv preprint arXiv:2103.14749, 2021 | |
| dc.relation.references | A. Uma, D. Almanea, and M. Poesio, “Scaling and disagreements: Bias, noise, and ambiguity,” Frontiers in Artificial Intelligence, vol. 5, p. 818451, 2022 | |
| dc.relation.references | N. Garcia, Y. Hirota, Y. Wu, and Y. Nakashima, “Uncurated image-text datasets: Shedding light on demographic bias,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6957–6966, 2023 | |
| dc.relation.references | L. Zhang, R. Tanno, M. Xu, Y. Huang, K. Bronik, C. Jin, J. Jacob, Y. Zheng, L. Shao, O. Ciccarelli, et al., “Learning from multiple annotators for medical image segmentation,” Pattern Recognition, vol. 138, p. 109400, 2023 | |
| dc.relation.references | M. H. M. Noor and A. O. Ige, “A survey on state-of-the-art deep learning applications and challenges,” Engineering Applications of Artificial Intelligence, vol. 159, p. 111225, 2025 | |
| dc.relation.references | X. Wang and W. Jia, “Optimizing edge ai: a comprehensive survey on data, model, and system strategies,” arXiv preprint arXiv:2501.03265, 2025 | |
| dc.relation.references | C. Silvano, D. Ielmini, F. Ferrandi, L. Fiorin, S. Curzel, L. Benini, F. Conti, A. Garofalo, C. Zambelli, E. Calore, et al., “A survey on deep learning hardware accelerators for heterogeneous hpc platforms,” ACM Computing Surveys, vol. 57, no. 11, pp. 1–39, 2025 | |
| dc.relation.references | S. Zhu, T. Yu, T. Xu, H. Chen, S. Dustdar, S. Gigan, D. Gunduz, E. Hossain, Y. Jin, F. Lin, et al., “Intelligent computing: the latest advances, challenges, and future,” Intelligent Computing, vol. 2, p. 0006, 2023 | |
| dc.relation.references | H.-I. Liu, M. Galindo, H. Xie, L.-K. Wong, H.-H. Shuai, Y.-H. Li, and W.-H. Cheng, “Lightweight deep learning for resource-constrained environments: A survey,” ACM Computing Surveys, vol. 56, no. 10, pp. 1–42, 2024 | |
| dc.relation.references | E. Edozie, A. N. Shuaibu, U. K. John, and B. O. Sadiq, “Comprehensive review of recent developments in visual object detection based on deep learning,” Artificial Intelligence Review, vol. 58, no. 9, p. 277, 2025 | |
| dc.relation.references | I. Mavromatis, K. Katsaros, and A. Khan, “Computing within limits: An empirical study of energy consumption in ml training and inference,” arXiv preprint arXiv:2406.14328, 2024 | |
| dc.relation.references | X. Wang, Z. Tang, J. Guo, T. Meng, C. Wang, T. Wang, and W. Jia, “Empowering edge intelligence: A comprehensive survey on on-device ai models,” ACM Computing Surveys, vol. 57, no. 9, pp. 1–39, 2025 | |
| dc.relation.references | P. Tsirtsakis, G. Zacharis, G. S. Maraslidis, and G. F. Fragulis, “Deep learning for object recognition: A comprehensive review of models and algorithms,” International Journal of Cognitive Computing in Engineering, 2025 | |
| dc.relation.references | A. Kamilaris and F. X. Prenafeta-Boldú, “Deep learning in agriculture: A survey,” Computers and electronics in agriculture, vol. 147, pp. 70–90, 2018 | |
| dc.relation.references | A. Khan, A. Sohail, U. Zahoora, and A. S. Qureshi, “A survey of the recent architectures of deep convolutional neural networks,” Artificial intelligence review, vol. 53, no. 8, pp. 5455–5516, 2020 | |
| dc.relation.references | I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning. MIT Press, 2016. http://www.deeplearningbook.org | |
| dc.relation.references | B. Barz and J. Denzler, “Deep learning on small datasets without pre-training using cosine loss,” in Proceedings of the IEEE/CVF winter conference on applications of computer vision, pp. 1371–1380, 2020 | |
| dc.relation.references | V. Cheplygina, M. De Bruijne, and J. P. Pluim, “Not-so-supervised: a survey of semi-supervised, multi-instance, and transfer learning in medical image analysis,” Medical image analysis, vol. 54, pp. 280–296, 2019 | |
| dc.relation.references | M. Z. Alom, T. M. Taha, C. Yakopcic, S. Westberg, P. Sidike, M. S. Nasrin, M. Hasan, B. C. Van Essen, A. A. Awwal, and V. K. Asari, “A state-of-the-art survey on deep learning theory and architectures,” electronics, vol. 8, no. 3, p. 292, 2019 | |
| dc.relation.references | W. Rawat and Z. Wang, “Deep convolutional neural networks for image classification: A comprehensive review,” Neural computation, vol. 29, no. 9, pp. 2352–2449, 2017 | |
| dc.relation.references | T.-Y. Lin, M. Maire, S. Belongie, J. Hays, P. Perona, D. Ramanan, P. Dollár, and C. L. Zitnick, “Microsoft coco: Common objects in context,” in European conference on computer vision, pp. 740–755, Springer, 2014 | |
| dc.relation.references | A. Esteva, A. Robicquet, B. Ramsundar, V. Kuleshov, M. DePristo, K. Chou, C. Cui, G. Corrado, S. Thrun, and J. Dean, “A guide to deep learning in healthcare,” Nature medicine, vol. 25, no. 1, pp. 24–29, 2019 | |
| dc.relation.references | G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, J. A. Van Der Laak, B. Van Ginneken, and C. I. Sánchez, “A survey on deep learning in medical image analysis,” Medical image analysis, vol. 42, pp. 60–88, 2017 | |
| dc.relation.references | P. Rajpurkar, J. Irvin, K. Zhu, B. Yang, H. Mehta, T. Duan, D. Ding, A. Bagul, C. Langlotz, K. Shpanskaya, et al., “Chexnet: Radiologist-level pneumonia detection on chest x-rays with deep learning,” arXiv preprint arXiv:1711.05225, 2017 | |
| dc.relation.references | O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in International Conference on Medical image computing and computer-assisted intervention, pp. 234–241, Springer, 2015 | |
| dc.relation.references | S. Graham, Q. D. Vu, S. E. A. Raza, A. Azam, Y. W. Tsang, J. T. Kwak, and N. Rajpoot, “Hover-net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images,” Medical image analysis, vol. 58, p. 101563, 2019 | |
| dc.relation.references | L. Luo, X. Wang, Y. Lin, X. Ma, A. Tan, R. Chan, V. Vardhanabhuti, W. C. Chu, K.-T. Cheng, and H. Chen, “Deep learning in breast cancer imaging: A decade of progress and future directions,” IEEE Reviews in Biomedical Engineering, 2024 | |
| dc.relation.references | M. Harouni, V. Goyal, G. Feldman, S. Michael, and T. C. Voss, “Deep multi-scale and attention-based architectures for semantic segmentation in biomedical imaging.,” Computers, Materials & Continua, vol. 85, no. 1, 2025 | |
| dc.relation.references | A. H. Ornek, M. Ceylan, and S. Ervural, “Health status detection of neonates using infrared thermography and deep convolutional neural networks,” Infrared Physics & Technology, vol. 103, p. 103044, 2019 | |
| dc.relation.references | A. A. Bruins, K. R. Kistemaker, A. Boom, J. H. Klaessens, R. M. Verdaasdonk, and C. Boer, “Thermographic skin temperature measurement compared with cold sensation in predicting the efficacy and distribution of epidural anesthesia,” Journal of clinical monitoring and computing, vol. 32, no. 2, pp. 335–341, 2018 | |
| dc.relation.references | T. Trongtirakul, K. Panetta, A. M. Grigoryan, and S. S. Agaian, “A novel entropy-based approach for thermal image segmentation using multilevel thresholding,” Entropy, vol. 27, no. 5, p. 526, 2025 | |
| dc.relation.references | L. Li, Q. Zhang, and D. Huang, “A review of imaging techniques for plant phenotyping,” Sensors, vol. 14, no. 11, pp. 20078–20111, 2014 | |
| dc.relation.references | S. P. Mohanty, D. P. Hughes, and M. Salathé, “Using deep learning for image-based plant disease detection,” Frontiers in Plant Science, vol. 7, 2016 | |
| dc.relation.references | D. C. Tsouros, S. Bibi, and P. G. Sarigiannidis, “A review on uav-based applications for precision agriculture,” Information, vol. 10, no. 11, p. 349, 2019 | |
| dc.relation.references | A. Upadhyay, N. S. Chandel, K. P. Singh, S. K. Chakraborty, B. M. Nandede, M. Kumar, A. Subeesh, K. Upendar, A. Salem, and A. Elbeltagi, “Deep learning and computer vision in plant disease detection: a comprehensive review of techniques, models, and trends in precision agriculture,” Artificial Intelligence Review, vol. 58, no. 3, p. 92, 2025 | |
| dc.relation.references | M. H. Tanveer, Z. Fatima, S. Zardari, and D. Guerra-Zubiaga, “An in-depth analysis of domain adaptation in computer and robotic vision,” Applied Sciences, vol. 13, no. 23, p. 12823, 2023 | |
| dc.relation.references | Y. Shi, L. Han, X. Zhang, T. Sobeih, T. Gaiser, N. H. Thuy, D. Behrend, A. K. Srivastava, K. Halder, and F. Ewert, “Deep learning meets process-based models: A hybrid approach to agricultural challenges,” arXiv preprint arXiv:2504.16141, 2025 | |
| dc.relation.references | H. B. Li, F. Navarro, I. Ezhov, A. Bayat, D. Das, F. Kofler, S. Shit, D. Waldmannstetter, J. C. Paetzold, X. Hu, et al., “Qubiq: Uncertainty quantification for biomedical image segmentation challenge,” arXiv preprint arXiv:2405.18435, 2024 | |
| dc.relation.references | D. Gurari, Q. Li, A. J. Stangl, A. Guo, C. Lin, K. Grauman, J. Luo, and J. P. Bigham, “Vizwiz grand challenge: Answering visual questions from blind people,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3608–3617, 2018 | |
| dc.relation.references | A. M. Davani, M. Díaz, and V. Prabhakaran, “Dealing with disagreements: Looking beyond the majority vote in subjective annotations,” Transactions of the Association for Computational Linguistics, vol. 10, pp. 92–110, 2022 | |
| dc.relation.references | A. J. Lee, Y. Cho, Y.-s. Shin, A. Kim, and H. Myung, “Vivid++: Vision for visibility dataset,” IEEE Robotics and Automation Letters, vol. 7, no. 3, pp. 6282–6289, 2022 | |
| dc.relation.references | Y. Yan, R. Rosales, G. Fung, R. Subramanian, and J. Dy, “Learning from multiple annotators with varying expertise,” Machine learning, vol. 95, no. 3, pp. 291–327, 2014 | |
| dc.relation.references | N. Miskin, G. C. Gaviola, R. Y. Huang, C. J. Kim, T. C. Lee, K. M. Small, G. G. Wieschhoff, and J. C. Mandell, “Intra-and intersubspecialty variability in lumbar spine mri interpretation: a multireader study comparing musculoskeletal radiologists and neuroradiologists,” Current Problems in Diagnostic Radiology, vol. 49, no. 3, pp. 182–187, 2020 | |
| dc.relation.references | M. Ji, K. Zhang, Q. Wu, and Z. Deng, “Multi-label learning for crop leaf diseases recognition and severity estimation based on convolutional neural networks,” Soft Computing, vol. 24, no. 20, pp. 15327–15340, 2020 | |
| dc.relation.references | A. Jungo, R. Meier, E. Ermis, M. Blatti-Moreno, E. Herrmann, R. Wiest, and M. Reyes, “On the effect of inter-observer variability for a reliable estimation of uncertainty of medical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, pp. 682–690, Springer, 2018 | |
| dc.relation.references | D. T. D. Ha, “Crowdsourcing for large-scale data labelling,” 2025 | |
| dc.relation.references | A. Sorokin and D. Forsyth, “Utility data annotation with amazon mechanical turk,” in 2008 IEEE computer society conference on computer vision and pattern recognition workshops, pp. 1–8, IEEE, 2008 | |
| dc.relation.references | V. C. Raykar, S. Yu, L. H. Zhao, G. H. Valadez, C. Florin, L. Bogoni, and L. Moy, “Learning from crowds.,” Journal of machine learning research, vol. 11, no. 4, 2010 | |
| dc.relation.references | S. Paun, B. Carpenter, J. Chamberlain, D. Hovy, U. Kruschwitz, and M. Poesio, “Comparing bayesian models of annotation,” Transactions of the Association for Computational Linguistics, vol. 6, pp. 571–585, 2018 | |
| dc.relation.references | S. Albarqouni, C. Baur, F. Achilles, V. Belagiannis, S. Demirci, and N. Navab, “Aggnet: deep learning from crowds for mitosis detection in breast cancer histology images,” IEEE transactions on medical imaging, vol. 35, no. 5, pp. 1313–1321, 2016 | |
| dc.relation.references | Y. Xu, V. Derricks, A. Earl, and D. Jurgens, “Modeling annotator disagreement with demographic-aware experts and synthetic perspectives,” 2025 | |
| dc.relation.references | Z. Chen, H. Sun, H. He, and P. Chen, “Learning from noisy crowd labels with logics,” in 2023 IEEE 39th International Conference on Data Engineering (ICDE), pp. 41–52, IEEE, 2023 | |
| dc.relation.references | J. Tu, G. Yu, J. Wang, C. Domeniconi, and X. Zhang, “Attention-aware answers of the crowd,” in Proceedings of the 2020 SIAM International Conference on Data Mining, pp. 451–459, SIAM, 2020 | |
| dc.relation.references | M. Dong, A. Yang, Z. Wang, D. Li, J. Yang, and R. Zhao, “Uncertainty-aware consistency learning for semi-supervised medical image segmentation,” Knowledge-Based Systems, vol. 309, p. 112890, 2025 | |
| dc.relation.references | E. Guo, Z. Wang, Z. Zhao, and L. Zhou, “Imbalanced medical image segmentation with pixel-dependent noisy labels,” IEEE Transactions on Medical Imaging, 2024 | |
| dc.relation.references | D. Karimi, H. Dou, S. K. Warfield, and A. Gholipour, “Deep learning with noisy labels: Exploring techniques and remedies in medical image analysis,” Medical image analysis, vol. 65, p. 101759, 2020 | |
| dc.relation.references | A. Ghosh, H. Kumar, and P. S. Sastry, “Robust loss functions under label noise for deep neural networks,” in Proceedings of the AAAI conference on artificial intelligence, vol. 31, 2017 | |
| dc.relation.references | Z. Zhang and M. Sabuncu, “Generalized cross entropy loss for training deep neural networks with noisy labels,” Advances in neural information processing systems, vol. 31, 2018 | |
| dc.relation.references | J. C. Triana-Martinez, J. Gil-González, J. A. Fernandez-Gallego, A. M. Álvarez-Meza, and C. G. Castellanos-Dominguez, “Chained deep learning using generalized cross-entropy for multiple annotators classification,” Sensors, vol. 23, no. 7, p. 3518, 2023 | |
| dc.relation.references | J. Gil-González, A. Valencia-Duque, A. Álvarez-Meza, Á. Orozco-Gutiérrez, and A. García-Moreno, “Regularized chained deep neural network classifier for multiple annotators,” Applied Sciences, vol. 11, no. 12, p. 5409, 2021 | |
| dc.relation.references | H. Wei, L. Feng, X. Chen, and B. An, “Combating noisy labels by agreement: A joint training method with co-regularization,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 13726–13735, 2020 | |
| dc.relation.references | B. Han, Q. Yao, X. Yu, G. Niu, M. Xu, W. Hu, I. Tsang, and M. Sugiyama, “Co-teaching: Robust training of deep neural networks with extremely noisy labels,” Advances in neural information processing systems, vol. 31, 2018 | |
| dc.relation.references | S. Reed, H. Lee, D. Anguelov, C. Szegedy, D. Erhan, and A. Rabinovich, “Training deep neural networks on noisy labels with bootstrapping,” arXiv preprint arXiv:1412.6596, 2014 | |
| dc.relation.references | M. Lukasik, S. Bhojanapalli, A. Menon, and S. Kumar, “Does label smoothing mitigate label noise?,” in International Conference on Machine Learning, pp. 6448–6458, PMLR, 2020 | |
| dc.relation.references | X. Qi, Z. Zhang, C. Gang, H. Zhang, L. Zhang, Z. Zhang, and Y. Zhao, “Mediaug: Exploring visual augmentation in medical imaging,” in Annual Conference on Medical Image Understanding and Analysis, pp. 218–232, Springer, 2025 | |
| dc.relation.references | A. Tupper and C. Gagné, “Analyzing data augmentation for medical images: A case study in ultrasound images,” arXiv preprint arXiv:2403.09828, 2024 | |
| dc.relation.references | S. Nesteruk, D. Shadrin, and M. Pukalchik, “Image augmentation for multitask few-shot learning: Agricultural domain use-case,” arXiv preprint arXiv:2102.12295, 2021 | |
| dc.relation.references | Z. Niu, S. Ouyang, S. Xie, Y. wei Chen, and L. Lin, “A survey on domain generalization for medical image analysis,” 2024 | |
| dc.relation.references | Y. Skandarani, P.-M. Jodoin, and A. Lalande, “Gans for medical image synthesis: An empirical study,” Journal of Imaging, vol. 9, no. 3, p. 69, 2023 | |
| dc.relation.references | Q. Dou, C. Ouyang, C. Chen, H. Chen, and P.-A. Heng, “Unsupervised cross-modality domain adaptation of convnets for biomedical image segmentations with adversarial loss,” arXiv preprint arXiv:1804.10916, 2018 | |
| dc.relation.references | J. Zhang, Y. Zheng, and Y. Shi, “A soft label method for medical image segmentation with multirater annotations,” Computational Intelligence and Neuroscience, vol. 2023, no. 1, p. 1883597, 2023 | |
| dc.relation.references | Y. Bi, B. Xue, P. Mesejo, S. Cagnoni, and M. Zhang, “A survey on evolutionary computation for computer vision and image analysis: Past, present, and future trends,” IEEE Transactions on Evolutionary Computation, vol. 27, p. 5–25, Feb. 2023 | |
| dc.relation.references | R. Yamashita, M. Nishio, R. K. G. Do, and K. Togashi, “Convolutional neural networks: an overview and application in radiology,” Insights into imaging, vol. 9, no. 4, pp. 611–629, 2018 | |
| dc.relation.references | L. Alzubaidi, J. Zhang, A. J. Humaidi, A. Al-Dujaili, Y. Duan, O. Al-Shamma, J. Santamaría, M. A. Fadhel, M. Al-Amidie, and L. Farhan, “Review of deep learning: concepts, cnn architectures, challenges, applications, future directions,” Journal of big Data, vol. 8, no. 1, p. 53, 2021 | |
| dc.relation.references | M. E. Rayed, S. S. Islam, S. I. Niha, J. R. Jim, M. M. Kabir, and M. Mridha, “Deep learning for medical image segmentation: State-of-the-art advancements and challenges,” Informatics in medicine unlocked, vol. 47, p. 101504, 2024 | |
| dc.relation.references | S. Takahashi, Y. Sakaguchi, N. Kouno, K. Takasawa, K. Ishizu, Y. Akagi, R. Aoyama, N. Teraya, A. Bolatkan, N. Shinkai, et al., “Comparison of vision transformers and convolutional neural networks in medical image analysis: A systematic review,” Journal of Medical Systems, vol. 48, no. 1, p. 84, 2024 | |
| dc.relation.references | D. E. Boukhari, “Mamba-cnn: A hybrid architecture for efficient and accurate facial beauty prediction,” 2025 | |
| dc.relation.references | G. A. Pereira and M. Hussain, “A review of transformer-based models for computer vision tasks: Capturing global context and spatial relationships,” 2024 | |
| dc.relation.references | A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020 | |
| dc.relation.references | N. Carion, F. Massa, G. Synnaeve, N. Usunier, A. Kirillov, and S. Zagoruyko, “End-to-end object detection with transformers,” in European conference on computer vision, pp. 213–229, Springer, 2020 | |
| dc.relation.references | S. Zheng, J. Lu, H. Zhao, X. Zhu, Z. Luo, Y. Wang, Y. Fu, J. Feng, T. Xiang, P. H. Torr, et al., “Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 6881–6890, 2021 | |
| dc.relation.references | Z. Peng, W. Huang, S. Gu, L. Xie, Y. Wang, J. Jiao, and Q. Ye, “Conformer: Local features coupling global representations for visual recognition,” in Proceedings of the IEEE/CVF international conference on computer vision, pp. 367–376, 2021 | |
| dc.relation.references | Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, and B. Guo, “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision, pp. 10012–10022, 2021 | |
| dc.relation.references | Z. Liu, H. Mao, C.-Y. Wu, C. Feichtenhofer, T. Darrell, and S. Xie, “A convnet for the 2020s,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 11976–11986, 2022 | |
| dc.relation.references | J. Pan, A. Bulat, F. Tan, X. Zhu, L. Dudziak, H. Li, G. Tzimiropoulos, and B. Martinez, “Edgevits: Competing light-weight cnns on mobile devices with vision transformers,” in European conference on computer vision, pp. 294–311, Springer, 2022 | |
| dc.relation.references | G. Xu, Z. Hao, Y. Luo, H. Hu, J. An, and S. Mao, “Devit: Decomposing vision transformers for collaborative inference in edge devices,” IEEE Transactions on Mobile Computing, vol. 23, no. 5, pp. 5917–5932, 2023 | |
| dc.relation.references | E. Strubell, A. Ganesh, and A. McCallum, “Energy and policy considerations for modern deep learning research,” in Proceedings of the AAAI conference on artificial intelligence, vol. 34, pp. 13693–13696, 2020 | |
| dc.relation.references | B. Jacob, S. Kligys, B. Chen, M. Zhu, M. Tang, A. Howard, H. Adam, and D. Kalenichenko, “Quantization and training of neural networks for efficient integer-arithmetic-only inference,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2704–2713, 2018 | |
| dc.relation.references | M. Tan and Q. Le, “Efficientnet: Rethinking model scaling for convolutional neural networks,” in International conference on machine learning, pp. 6105–6114, PMLR, 2019 | |
| dc.relation.references | P. Mittal, “A comprehensive survey of deep learning-based lightweight object detection models for edge devices,” Artificial Intelligence Review, vol. 57, no. 9, p. 242, 2024 | |
| dc.relation.references | J. Chen and X. Ran, “Deep learning with edge computing: A review,” Proceedings of the IEEE, vol. 107, no. 8, pp. 1655–1674, 2019 | |
| dc.relation.references | M. Tan and Q. Le, “Efficientnetv2: Smaller models and faster training,” in International conference on machine learning, pp. 10096–10106, PMLR, 2021 S. Han, H. Mao, and W. J. Dally, “Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding,” arXiv preprint arXiv:1510.00149, 2015 Y. He, X. Zhang, and J. Sun, “Channel pruning for accelerating very deep neural networks,” in Proceedings of the IEEE international conference on computer vision, pp. 1389–1397, 2017 G. Hinton, O. Vinyals, and J. Dean, “Distilling the knowledge in a neural network,” arXiv preprint arXiv:1503.02531, 2015 N. Lee, T. Ajanthan, and P. H. Torr, “Snip: Single-shot network pruning based on connection sensitivity,” arXiv preprint arXiv:1810.02340, 2018 S. Mehta and M. Rastegari, “Mobilevit: light-weight, general-purpose, and mobile-friendly vision transformer,” arXiv preprint arXiv:2110.02178, 2021 K. Choromanski, V. Likhosherstov, D. Dohan, X. Song, A. Gane, T. Sarlos, P. Hawkins, J. Davis, A. Mohiuddin, L. Kaiser, et al., “Rethinking attention with performers,” arXiv preprint arXiv:2009.14794, 2020 X. Min, Y. Ye, S. Xiong, and X. Chen, “Computer vision meets generative models in agriculture: Technological advances, challenges and opportunities.,” Applied Sciences (2076-3417), vol. 15, no. 14, 2025 W. Lei, H. Wang, R. Gu, S. Zhang, S. Zhang, and G. Wang, “Deepigeos-v2: Deep interactive segmentation of multiple organs from head and neck images with lightweight cnns,” in International Workshop on Large-scale Annotation of Biomedical data and Expert Label Synthesis, pp. 61–69, Springer, 2019 X. Zhang, Z. Cao, and W. Dong, “Overview of edge computing in the agricultural internet of things: Key technologies, applications, challenges,” Ieee Access, vol. 8, pp. 141748–141761, 2020 F. Wang, M. Zhang, X. Wang, X. Ma, and J. Liu, “Deep learning for edge computing applications: A state-of-the-art survey,” IEEE Access, vol. 8, pp. 58322–58336, 2020 L. Bouza, A. Bugeau, and L. Lannelongue, “How to estimate carbon footprint when training deep learning models? a guide and review,” Environmental Research Communications, vol. 5, no. 11, p. 115014, 2023 D. Patterson, J. Gonzalez, Q. Le, C. Liang, L.-M. Munguia, D. Rothchild, D. So, M. Texier, and J. Dean, “Carbon emissions and large neural network training,” arXiv preprint arXiv:2104.10350, 2021 L. H. Kaack, P. L. Donti, E. Strubell, G. Kamiya, F. Creutzig, and D. Rolnick, “Aligning artificial intelligence with climate change mitigation,” Nature Climate Change, vol. 12, no. 6, pp. 518–527, 2022 K.Han, Y.Wang, H.Chen, X.Chen, J. Guo, Z. Liu, Y. Tang, A. Xiao, C. Xu, Y. Xu, et al., “A survey on vision transformer,” IEEE transactions on pattern analysis and machine intelligence, vol. 45, no. 1, pp. 87–110, 2022 B. Graham, A. El-Nouby, H. Touvron, P. Stock, A. Joulin, H. Jégou, and M. Douze, “Levit: a vision transformer in convnet’s clothing for faster inference,” in Proceedings of the IEEE/CVF international conference on computer vision, pp. 12259–12269, 2021 C. Fan, Q. Su, Z. Xiao, H. Su, A. Hou, and B. Luan, “Vit-frd: A vision transformer model for cardiac mri image segmentation based on feature recombination distillation,” IEEE Access, vol. 11, pp. 129763–129772, 2023 J. Fan, B. Gao, Q. Ge, Y. Ran, J. Zhang, and H. Chu, “Segtransconv: Transformer and cnn hybrid method for real-time semantic segmentation of autonomous vehicles,” IEEE Transactions on Intelligent Transportation Systems, vol. 25, no. 2, pp. 1586–1601, 2023 O. M. Parkhi, A. Vedaldi, A. Zisserman, and C. Jawahar, “Cats and dogs,” in 2012 IEEE conference on computer vision and pattern recognition, pp. 3498–3505, IEEE, 2012 M. Amgad, H. Elfandy, H. Hussein, L. A. Atteya, M. A. Elsebaie, L. S. Abo Elnasr, R. A. Sakr, H. S. Salem, A. F. Ismail, A. M. Saad, et al., “Structured crowdsourcing enables convolutional segmentation of histology images,” Bioinformatics, vol. 35, no. 18, pp. 3461–3467, 2019 M. López-Pérez, P. Morales-Álvarez, L. A. Cooper, C. Felicelli, J. Goldstein, B. Vadasz, R. Molina, and A. K. Katsaggelos, “Learning from crowds for automated histopathological image segmentation,” Computerized Medical Imaging and Graphics, vol. 112, p. 102327, 2024 R. Mejia-Zuluaga, J. C. Aguirre-Arango, D. Collazos-Huertas, J. Daza-Castillo, N. Valencia-Marulanda, M. Calderón-Marulanda, Ó. Aguirre Ospina, A. Alvarez-Meza, and G. Castellanos-Dominguez, “Deep learning semantic segmentation of feet using infrared thermal images,” in Ibero-American Conference on Artificial Intelligence, pp. 342–352, Springer, 2022 J. C. Aguirre-Arango, A. M. Álvarez-Meza, and G. Castellanos-Dominguez, “Feet segmentation for regional analgesia monitoring using convolutional rff and layer-wise weighted cam interpretability,” Computation, vol. 11, no. 6, p. 113, 2023 | |
| dc.relation.references | S. Han, H. Mao, and W. J. Dally, “Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding,” arXiv preprint arXiv:1510.00149, 2015 | |
| dc.relation.references | Y. He, X. Zhang, and J. Sun, “Channel pruning for accelerating very deep neural networks,” in Proceedings of the IEEE international conference on computer vision, pp. 1389–1397, 2017 | |
| dc.relation.references | G. Hinton, O. Vinyals, and J. Dean, “Distilling the knowledge in a neural network,” arXiv preprint arXiv:1503.02531, 2015 | |
| dc.relation.references | N. Lee, T. Ajanthan, and P. H. Torr, “Snip: Single-shot network pruning based on connection sensitivity,” arXiv preprint arXiv:1810.02340, 2018 | |
| dc.relation.references | S. Mehta and M. Rastegari, “Mobilevit: light-weight, general-purpose, and mobile-friendly vision transformer,” arXiv preprint arXiv:2110.02178, 2021 | |
| dc.relation.references | K. Choromanski, V. Likhosherstov, D. Dohan, X. Song, A. Gane, T. Sarlos, P. Hawkins, J. Davis, A. Mohiuddin, L. Kaiser, et al., “Rethinking attention with performers,” arXiv preprint arXiv:2009.14794, 2020 | |
| dc.relation.references | X. Min, Y. Ye, S. Xiong, and X. Chen, “Computer vision meets generative models in agriculture: Technological advances, challenges and opportunities.,” Applied Sciences (2076-3417), vol. 15, no. 14, 2025 | |
| dc.relation.references | W. Lei, H. Wang, R. Gu, S. Zhang, S. Zhang, and G. Wang, “Deepigeos-v2: Deep interactive segmentation of multiple organs from head and neck images with lightweight cnns,” in International Workshop on Large-scale Annotation of Biomedical data and Expert Label Synthesis, pp. 61–69, Springer, 2019 | |
| dc.relation.references | X. Zhang, Z. Cao, and W. Dong, “Overview of edge computing in the agricultural internet of things: Key technologies, applications, challenges,” Ieee Access, vol. 8, pp. 141748–141761, 2020 | |
| dc.relation.references | F. Wang, M. Zhang, X. Wang, X. Ma, and J. Liu, “Deep learning for edge computing applications: A state-of-the-art survey,” IEEE Access, vol. 8, pp. 58322–58336, 2020 | |
| dc.relation.references | L. Bouza, A. Bugeau, and L. Lannelongue, “How to estimate carbon footprint when training deep learning models? a guide and review,” Environmental Research Communications, vol. 5, no. 11, p. 115014, 2023 | |
| dc.relation.references | D. Patterson, J. Gonzalez, Q. Le, C. Liang, L.-M. Munguia, D. Rothchild, D. So, M. Texier, and J. Dean, “Carbon emissions and large neural network training,” arXiv preprint arXiv:2104.10350, 2021 | |
| dc.relation.references | L. H. Kaack, P. L. Donti, E. Strubell, G. Kamiya, F. Creutzig, and D. Rolnick, “Aligning artificial intelligence with climate change mitigation,” Nature Climate Change, vol. 12, no. 6, pp. 518–527, 2022 | |
| dc.relation.references | K. Han, Y. Wang, H. Chen, X. Chen, J. Guo, Z. Liu, Y. Tang, A. Xiao, C. Xu, Y. Xu, et al., “A survey on vision transformer,” IEEE transactions on pattern analysis and machine intelligence, vol. 45, no. 1, pp. 87–110, 2022 | |
| dc.relation.references | B. Graham, A. El-Nouby, H. Touvron, P. Stock, A. Joulin, H. Jégou, and M. Douze, “Levit: a vision transformer in convnet’s clothing for faster inference,” in Proceedings of the IEEE/CVF international conference on computer vision, pp. 12259–12269, 2021 | |
| dc.relation.references | C. Fan, Q. Su, Z. Xiao, H. Su, A. Hou, and B. Luan, “Vit-frd: A vision transformer model for cardiac mri image segmentation based on feature recombination distillation,” IEEE Access, vol. 11, pp. 129763–129772, 2023 | |
| dc.relation.references | J. Fan, B. Gao, Q. Ge, Y. Ran, J. Zhang, and H. Chu, “Segtransconv: Transformer and cnn hybrid method for real-time semantic segmentation of autonomous vehicles,” IEEE Transactions on Intelligent Transportation Systems, vol. 25, no. 2, pp. 1586–1601, 2023 | |
| dc.relation.references | O. M. Parkhi, A. Vedaldi, A. Zisserman, and C. Jawahar, “Cats and dogs,” in 2012 IEEE conference on computer vision and pattern recognition, pp. 3498–3505, IEEE, 2012 | |
| dc.relation.references | M. Amgad, H. Elfandy, H. Hussein, L. A. Atteya, M. A. Elsebaie, L. S. Abo Elnasr, R. A. Sakr, H. S. Salem, A. F. Ismail, A. M. Saad, et al., “Structured crowdsourcing enables convolutional segmentation of histology images,” Bioinformatics, vol. 35, no. 18, pp. 3461–3467, 2019 | |
| dc.relation.references | M. López-Pérez, P. Morales-Álvarez, L. A. Cooper, C. Felicelli, J. Goldstein, B. Vadasz, R. Molina, and A. K. Katsaggelos, “Learning from crowds for automated histopathological image segmentation,” Computerized Medical Imaging and Graphics, vol. 112, p. 102327, 2024 | |
| dc.relation.references | R. Mejia-Zuluaga, J. C. Aguirre-Arango, D. Collazos-Huertas, J. Daza-Castillo, N. Valencia-Marulanda, M. Calderón-Marulanda, Ó. Aguirre Ospina, A. Alvarez-Meza, and G. Castellanos-Dominguez, “Deep learning semantic segmentation of feet using infrared thermal images,” in Ibero-American Conference on Artificial Intelligence, pp. 342–352, Springer, 2022 | |
| dc.relation.references | J. C. Aguirre-Arango, A. M. Álvarez-Meza, and G. Castellanos-Dominguez, “Feet segmentation for regional analgesia monitoring using convolutional rff and layer-wise weighted cam interpretability,” Computation, vol. 11, no. 6, p. 113, 2023 | |
| dc.relation.references | M. Abdelrahman, H. Abdel-Haleem, M. El-Sayed, E. S. Abdel-Razik, S. Jogaiah, and D. J. Burritt, “Exogenously applied proline and salicylic acid alleviate the harmful effects of osmotic stress on tomato seed germination and seedlings growth,” Agronomy, vol. 11, no. 1, p. 5, 2021 | |
| dc.relation.references | Roboflow, Inc., “Roboflow: Computer vision development platform.” https://roboflow.com, 2025 | |
| dc.relation.references | K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770–778, 2016 | |
| dc.relation.references | F. Isensee, P. F. Jaeger, S. A. Kohl, J. Petersen, and K. H. Maier-Hein, “nnu-net: a self-configuring method for deep learning-based biomedical image segmentation,” Nature methods, vol. 18, no. 2, pp. 203–211, 2021 | |
| dc.relation.references | J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 3431–3440, 2015 | |
| dc.relation.references | I. D. Mienye and T. G. Swart, “A comprehensive review of deep learning: Architectures, recent advances, and applications,” Information, vol. 15, no. 12, p. 755, 2024 | |
| dc.relation.references | F. Rosenblatt, “The perceptron: a probabilistic model for information storage and organization in the brain.,” Psychological review, vol. 65, no. 6, p. 386, 1958 | |
| dc.relation.references | M. Minsky and S. Papert, “An introduction to computational geometry,” Cambridge tiass., HIT, vol. 479, no. 480, p. 104, 1969 | |
| dc.relation.references | K. Hornik, M. Stinchcombe, and H. White, “Multilayer feedforward networks are universal approximators,” Neural networks, vol. 2, no. 5, pp. 359–366, 1989 | |
| dc.relation.references | Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. Hubbard, and L. D. Jackel, “Backpropagation applied to handwritten zip code recognition,” Neural computation, vol. 1, no. 4, pp. 541–551, 1989 | |
| dc.relation.references | A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in neural information processing systems, vol. 25, 2012 | |
| dc.relation.references | S. Minaee, Y. Boykov, F. Porikli, A. Plaza, N. Kehtarnavaz, and D. Terzopoulos, “Image segmentation using deep learning: A survey,” IEEE transactions on pattern analysis and machine intelligence, vol. 44, no. 7, pp. 3523–3542, 2021 | |
| dc.relation.references | L.-C. Chen, G. Papandreou, F. Schroff, and H. Adam, “Rethinking atrous convolution for semantic image segmentation,” arXiv preprint arXiv:1706.05587, 2017 | |
| dc.relation.references | E. Kussul, T. Baidyk, L. Kasatkina, and V. Lukovich, “Rosenblatt perceptrons for handwritten digit recognition,” in IJCNN’01. International Joint Conference on Neural Networks. Proceedings (Cat. No. 01CH37222), vol. 2, pp. 1516–1520, IEEE, 2001 | |
| dc.relation.references | M. Marvin and A. P. Seymour, “Perceptrons,” Cambridge, MA: MIT Press, vol. 6, no. 318-362, p. 7, 1969 | |
| dc.relation.references | M. L. Minsky and S. A. Papert, “Perceptrons: expanded edition,” 1988 | |
| dc.relation.references | A. J. Al-Mahasneh, S. G. Anavatti, and M. A. Garratt, “The development of neural networks applications from perceptron to deep learning,” in 2017 International Conference on Advanced Mechatronics, Intelligent Manufacture, and Industrial Automation (ICAMIMIA), pp. 1–6, IEEE, 2017 | |
| dc.relation.references | H. Peters, “Perceptrons,” in Artificial Neural Networks: An Introduction to ANN Theory and Practice, pp. 67–81, Springer, 2005 | |
| dc.relation.references | P. E. Hart, D. G. Stork, and R. Duda, Pattern classification. Wiley Hoboken, 2001 | |
| dc.relation.references | R. Khardon and G. Wachman, “Noise tolerant variants of the perceptron algorithm.,” Journal of Machine Learning Research, vol. 8, no. 2, 2007 | |
| dc.relation.references | D. E. Rumelhart, G. E. Hinton, and R. J. Williams, “Learning representations by back-propagating errors,” nature, vol. 323, no. 6088, pp. 533–536, 1986 | |
| dc.relation.references | R. J. Williams, ““learning representations by back-propagating errors david e. rumelhart, geoffrey e. hinton, and,” Cognitive Modeling, p. 213, 2002 | |
| dc.relation.references | S.-i. Amari, “Backpropagation and stochastic gradient descent method,” Neurocomputing, vol. 5, no. 4-5, pp. 185–196, 1993 | |
| dc.relation.references | Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 2002 | |
| dc.relation.references | K. Hornik, “Approximation capabilities of multilayer feedforward networks,” Neural networks, vol. 4, no. 2, pp. 251–257, 1991 | |
| dc.relation.references | G. Bachmann, S. Anagnostidis, and T. Hofmann, “Scaling mlps: A tale of inductive bias,” Advances in Neural Information Processing Systems, vol. 36, pp. 60821–60840, 2023 | |
| dc.relation.references | I. O. Tolstikhin, N. Houlsby, A. Kolesnikov, L. Beyer, X. Zhai, T. Unterthiner, J. Yung, A. Steiner, D. Keysers, J. Uszkoreit, et al., “Mlp-mixer: An all-mlp architecture for vision,” Advances in neural information processing systems, vol. 34, pp. 24261–24272, 2021 | |
| dc.relation.references | K. Han, Y. Wang, Q. Tian, J. Guo, C. Xu, and C. Xu, “Ghostnet: More features from cheap operations,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 1580–1589, 2020 | |
| dc.relation.references | L.-C. Chen, Y. Zhu, G. Papandreou, F. Schroff, and H. Adam, “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European conference on computer vision (ECCV), pp. 801–818, 2018 | |
| dc.relation.references | E. Karakullukcu, “Leveraging convolutional neural networks for image-based classification of feature matrix data,” Expert Systems with Applications, vol. 281, p. 127625, 2025 | |
| dc.relation.references | C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, “Rethinking the inception architecture for computer vision,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 2818–2826, 2016 | |
| dc.relation.references | X. Glorot and Y. Bengio, “Understanding the difficulty of training deep feedforward neural networks,” in Proceedings of the thirteenth international conference on artificial intelligence and statistics, pp. 249–256, JMLR Workshop and Conference Proceedings, 2010 | |
| dc.relation.references | H. R. Roth, H. Oda, X. Zhou, N. Shimizu, Y. Yang, Y. Hayashi, M. Oda, M. Fujiwara, K. Misawa, and K. Mori, “An application of cascaded 3d fully convolutional networks for medical image segmentation,” Computerized Medical Imaging and Graphics, vol. 66, pp. 90–99, 2018 | |
| dc.relation.references | X. Xu, Q. Lu, L. Yang, S. Hu, D. Chen, Y. Hu, and Y. Shi, “Quantization of fully convolutional networks for accurate biomedical image segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 8300–8308, 2018 | |
| dc.relation.references | J. Hestness, S. Narang, N. Ardalani, G. Diamos, H. Jun, H. Kianinejad, M. M. A. Patwary, Y. Yang, and Y. Zhou, “Deep learning scaling is predictable, empirically,” arXiv preprint arXiv:1712.00409, 2017 | |
| dc.relation.references | A. Younesi, M. Ansari, M. Fazli, A. Ejlali, M. Shafique, and J. Henkel, “A comprehensive survey of convolutions in deep learning: Applications, challenges, and future trends,” IEEE Access, vol. 12, pp. 41180–41218, 2024 | |
| dc.relation.references | P. O.Pinheiro and R. Collobert, “From image-level to pixel-level labeling with convolutional networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1713–1721, 2015 | |
| dc.relation.references | Z. Zhang, Q. Liu, and Y. Wang, “Road extraction by deep residual u-net,” IEEE Geoscience and Remote Sensing Letters, vol. 15, no. 5, pp. 749–753, 2018 | |
| dc.relation.references | L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, and A. L. Yuille, “Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs,” IEEE transactions on pattern analysis and machine intelligence, vol. 40, no. 4, pp. 834–848, 2017 | |
| dc.relation.references | E. Shelhamer, J. Long, and T. Darrell, “Fully convolutional networks for semantic segmentation,” IEEE transactions on pattern analysis and machine intelligence, vol. 39, no. 4, pp. 640–651, 2017 | |
| dc.relation.references | H. Noh, S. Hong, and B. Han, “Learning deconvolution network for semantic segmentation,” in Proceedings of the IEEE international conference on computer vision, pp. 1520–1528, 2015 | |
| dc.relation.references | M. Drozdzal, E. Vorontsov, G. Chartrand, S. Kadoury, and C. Pal, “The importance of skip connections in biomedical image segmentation,” in International workshop on deep learning in medical image analysis, pp. 179–187, Springer, 2016 | |
| dc.relation.references | J. Fu, J. Liu, Y. Li, Y. Bao, W. Yan, Z. Fang, and H. Lu, “Contextual deconvolution network for semantic segmentation,” Pattern Recognition, vol. 101, p. 107152, 2020 | |
| dc.relation.references | X. Ren, Z. Deng, J. Ye, J. He, and D. Yang, “Fcn+: Global receptive convolution makes fcn great again,” Neurocomputing, vol. 631, p. 129655, 2025 | |
| dc.relation.references | Y. Yuan and Y. Cheng, “Medical image segmentation with unet-based multi-scale context fusion,” Scientific Reports, vol. 14, no. 1, p. 15687, 2024 | |
| dc.relation.references | N. Siddique, P. Sidike, C. Elkin, and V. Devabhaktuni, “U-net and its variants for medical image segmentation: theory and applications,” arXiv preprint arXiv:2011.01118, 2020 | |
| dc.relation.references | Z. Tang, X. Peng, K. Li, and D. N. Metaxas, “Towards efficient u-nets: A coupled and quantized approach,” IEEE transactions on pattern analysis and machine intelligence, vol. 42, no. 8, pp. 2038–2050, 2019 | |
| dc.relation.references | X. Xia and B. Kulis, “W-net: A deep model for fully unsupervised image segmentation,” arXiv preprint arXiv:1711.08506, 2017 | |
| dc.relation.references | H. Cao,Y. Wang,J. Chen, D. Jiang, X. Zhang, Q. Tian, and M. Wang, “Swin-unet: Unet-likepuretransformer for medical image segmentation,” in European conference on computer vision, pp. 205–218, Springer, 2022 | |
| dc.relation.references | R. Azad, E. K. Aghdam, A. Rauland, Y. Jia, A. H. Avval, A. Bozorgpour, S. Karimijafarbigloo, J. P. Cohen, E. Adeli, and D. Merhof, “Medical image segmentation review: The success of u-net,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024 | |
| dc.relation.references | M. Z. Alom, M. Hasan, C. Yakopcic, T. M. Taha, and V. K. Asari, “Recurrent residual convolutional neural network based on u-net (r2u-net) for medical image segmentation,” arXiv preprint arXiv:1802.06955, 2018 | |
| dc.relation.references | D. Jha, P. H. Smedsrud, M. A. Riegler, D. Johansen, T. De Lange, P. Halvorsen, and H. D. Johansen, “Resunet++: An advanced architecture for medical image segmentation,” in 2019 IEEE international symposium on multimedia (ISM), pp. 225–2255, IEEE, 2019 | |
| dc.relation.references | N. Siddique, S. Paheding, M. Z. Alom, and V. Devabhaktuni, “Recurrent residual u-net with efficientnet encoder for medical image segmentation,” in Pattern Recognition and Tracking XXXII, vol. 11735, pp. 134–142, SPIE, 2021 | |
| dc.relation.references | S. Metlek, “Cellsegunet: an improved deep segmentation model for the cell segmentation based on unet++ and residual unet models,” Neural Computing and Applications, vol. 36, no. 11, pp. 5799–5825, 2024 | |
| dc.relation.references | M. Z. Alom, C. Yakopcic, M. Hasan, T. M. Taha, and V. K. Asari, “Recurrent residual u-net for medical image segmentation,” Journal of medical imaging, vol. 6, no. 1, pp. 014006–014006, 2019 | |
| dc.relation.references | C. Liu, L.-C. Chen, F. Schroff, H. Adam, W. Hua, A. L. Yuille, and L. Fei-Fei, “Auto-deeplab: Hierarchical neural architecture search for semantic image segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp. 82–92, 2019 | |
| dc.relation.references | S. C. Yurtkulu, Y. H. Şahin, and G. Unal, “Semantic segmentation with extended deeplabv3 architecture,” in 2019 27th signal processing and communications applications conference (SIU), pp. 1–4, IEEE, 2019 | |
| dc.relation.references | S. Du, S. Du, B. Liu, and X. Zhang, “Incorporating deeplabv3+ and object-based image analysis for semantic segmentation of very high resolution remote sensing images,” International Journal of Digital Earth, vol. 14, no. 3, pp. 357–378, 2021 | |
| dc.relation.references | P. Ding and H. Qian, “Light-deeplabv3+: a lightweight real-time semantic segmentation method for complex environment perception,” Journal of Real-Time Image Processing, vol. 21, no. 1, p. 1, 2024 | |
| dc.relation.references | L.-C. Chen and Y. Zhu, “Semantic image segmentation with deeplab in tensorflow,” Google AI Blog, 2019 | |
| dc.relation.references | J. Wang, X. Zhang, T. Yan, and A. Tan, “Dpnet: Dual-pyramid semantic segmentation network based on improved deeplabv3 plus,” Electronics, vol. 12, no. 14, p. 3161, 2023 | |
| dc.relation.references | P. Wang, P. Chen, Y. Yuan, D. Liu, Z. Huang, X. Hou, and G. Cottrell, “Understanding convolution for semantic segmentation,” in 2018 IEEE winter conference on applications of computer vision (WACV), pp. 1451–1460, Ieee, 2018 | |
| dc.relation.references | S. Jadon, “A survey of loss functions for semantic segmentation,” 2020 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), pp. 1–7, 2020 | |
| dc.relation.references | A. A. Taha and A. Hanbury, “Metrics for evaluating 3d medical image segmentation: analysis, selection, and tool,” BMC Medical Imaging, vol. 15, no. 1, p. 29, 2015 | |
| dc.relation.references | H. Rezatofighi, N. Tsoi, J. Gwak, A. Sadeghian, I. Reid, and S. Savarese, “Generalized intersection over union: A metric and a loss for bounding box regression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 658–666, 2019 | |
| dc.relation.references | A. Garcia-Garcia, S. Orts-Escolano, S. Oprea, V. Villena-Martinez, and J. Garcia-Rodriguez, “A review on deep learning techniques applied to semantic segmentation,” Applied Soft Computing, vol. 70, pp. 41–65, 2018 | |
| dc.relation.references | O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 234–241, 2015 | |
| dc.relation.references | F. Milletari, N. Navab, and S.-A. Ahmadi, “V-net: Fully convolutional neural networks for volumetric medical image segmentation,” in 2016 Fourth International Conference on 3D Vision (3DV), pp. 565–571, 2016 | |
| dc.relation.references | S. S. M. Salehi, D. Erdogmus, and A. Gholipour, “Tversky loss function for image segmentation using 3d fully convolutional deep networks,” in International workshop on machine learning in medical imaging, pp. 379–387, Springer, 2017 | |
| dc.relation.references | T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollár, “Focal loss for dense object detection,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2980–2988, 2017 | |
| dc.relation.references | V. Yeghiazaryan and I. Voiculescu, “An overview of current evaluation methods used in medical image segmentation,” arXiv preprint arXiv:1512.01881, 2018 | |
| dc.relation.references | L. Maier-Hein, M. Eisenmann, A. Reinke, S. Onogur, M. Stankovic, P. Scholz, et al., “Why rankings of biomedical image analysis competitions should be interpreted with care,” Nature Communications, vol. 9, no. 1, p. 5217, 2018 | |
| dc.relation.references | T. Saito and M. Rehmsmeier, “The precision-recall plot is more informative than the roc plot when evaluating binary classifiers on imbalanced datasets,” in PLOS ONE, vol. 10, p. e0118432, 2015 | |
| dc.relation.references | A. Mao, M. Mohri, and Y. Zhong, “Cross-entropy loss functions: Theoretical analysis and applications,” in International conference on Machine learning, pp. 23803–23828, pmlr, 2023 | |
| dc.relation.references | R. Zhao, B. Qian, X. Zhang, Y. Li, R. Wei, Y. Liu, and Y. Pan, “Rethinking dice loss for medical image segmentation,” in 2020 IEEE international conference on data mining (ICDM), pp. 851–860, Ieee, 2020 | |
| dc.relation.references | A. W. Setiawan, “Image segmentation metrics in skin lesion: accuracy, sensitivity, specificity, dice coefficient, jaccard index, and matthews correlation coefficient,” in 2020 International Conference on Computer Engineering, Network, and Intelligent Multimedia (CENIM), pp. 97–102, IEEE, 2020 | |
| dc.relation.references | D. Müller, I. Soto-Rey, and F. Kramer, “Towards a guideline for evaluation metrics in medical image segmentation,” BMC Research Notes, vol. 15, no. 1, p. 210, 2022 | |
| dc.relation.references | R. Usamentiaga, D. G. Lema, O. D. Pedrayes, and D. F. Garcia, “Automated surface defect detection in metals: a comparative review of object detection and semantic segmentation using deep learning,” IEEE Transactions on Industry Applications, vol. 58, no. 3, pp. 4203–4213, 2022 | |
| dc.relation.references | R. Khanam, M. Hussain, R. Hill, and P. Allen, “A comprehensive review of convolutional neural networks for defect detection in industrial applications,” IEEE Access, 2024 | |
| dc.relation.references | K. P. Murphy, Probabilistic machine learning: an introduction. MIT press, 2022 | |
| dc.relation.references | C. G. Northcutt, L. Jiang, and I. L. Chuang, “Confident learning: Estimating uncertainty in dataset labels,” Journal of Artificial Intelligence Research, vol. 70, pp. 1373–1411, 2021 | |
| dc.relation.references | R. McKinley, M. Rebsamen, R. Meier, and R. Wiest, “Uncertainty visualization of multiple expert segmentations in medical images,” in Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 650–658, 2020 | |
| dc.relation.references | Z. Ji, P. Cui, Y. Li, and B. Li, “Learning from noisy labels with structured graph regularization,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 4, pp. 1257–1271, 2021 | |
| dc.relation.references | C. Jensen, C. Albers, A. Lenkoski, and W. Vanpaemel, “Bayesian approaches to modeling multiple annotators,” Psychometrika, vol. 84, no. 3, pp. 665–692, 2019 | |
| dc.relation.references | A. P. Dawid and A. M. Skene, “Maximum likelihood estimation of observer error-rates using the em algorithm,” Journal of the Royal Statistical Society: Series C (Applied Statistics), vol. 28, no. 1, pp. 20–28, 1979 | |
| dc.relation.references | S. K. Warfield, K. H. Zou, and W. M. Wells, “Simultaneous truth and performance level estimation (staple): an algorithm for the validation of image segmentation,” IEEE Transactions on Medical Imaging, vol. 23, no. 7, pp. 903–921, 2004 | |
| dc.relation.references | A. J. Asman and B. A. Landman, “Robust statistical estimation of segmentation uncertainty using random walks: application to hippocampal segmentation,” Medical Image Analysis, vol. 15, no. 4, pp. 482–493, 2011 | |
| dc.relation.references | A. A. Taha and A. Hanbury, “Formal methods for objective comparison of segmentation variants,” Medical Image Analysis, vol. 34, pp. 142–154, 2017 | |
| dc.relation.references | C. Zhang, P. Patras, and H. Haddadi, “Edge intelligence: On-demand deep learning model co-inference with device–edge synergy,” Proceedings of the 26th Annual International Conference on Mobile Computing and Networking, 2020 | |
| dc.relation.references | R. P. Foundation, “Raspberry pi as a platform for edge ai.” https://www.raspberrypi.com/news/raspberry-pi-and-ai-at-the-edge/, 2021 | |
| dc.relation.references | N. D. Lane, P. Georgiev, and L. Qendro, “Wearable computing for health and fitness: Exploring the state of the art,” IEEE Pervasive Computing, vol. 13, no. 2, pp. 49–56, 2014 | |
| dc.relation.references | V. Sze, Y.-H. Chen, T.-J. Yang, and J. S. Emer, “Efficient processing of deep neural networks: A tutorial and survey,” Proceedings of the IEEE, vol. 105, no. 12, pp. 2295–2329, 2017 | |
| dc.relation.references | T. Authors, “Tensorflow lite: Deploying machine learning models on mobile and iot devices,” in Proceedings of the Workshop on Mobile Machine Learning (MLSys), 2019 | |
| dc.relation.references | R. David, P. Warden, D. Situnayake, et al., “Tensorflow lite micro: Embedded machine learning for tinyml systems,” Proceedings of Machine Learning and Systems, vol. 3, pp. 800–811, 2021 | |
| dc.relation.references | A. Ratner, S. H. Bach, H. Ehrenberg, J. Fries, S. Wu, and C. Ré, “Snorkel: Rapid training data creation with weak supervision,” in Proceedings of the VLDB endowment. International conference on very large data bases, vol. 11, p. 269, 2017 | |
| dc.relation.references | N. Griffioen, N. Rankovic, F. Zamberlan, and M. Punith, “Efficient annotation reduction with active learning for computer vision-based retail product recognition,” Journal of Computational Social Science, vol. 7, no. 1, pp. 1039–1070, 2024 | |
| dc.relation.references | K. Zhou, Z. Liu, X. Zhai, C. Li, and K. Saenko, “Guest editorial: Special issue on the promises and dangers of large vision models,” International Journal of Computer Vision, vol. 132, no. 4, pp. 1009–1011, 2024 | |
| dc.relation.references | K. Yang, K. Qinami, L. Fei-Fei, J. Deng, and O. Russakovsky, “Towards fairer datasets: Filtering and balancing the distribution of the people subtree in the imagenet hierarchy,” in Proceedings of the 2020 conference on fairness, accountability, and transparency, pp. 547–558, 2020 | |
| dc.relation.references | A. Torralba and A. A. Efros, “Unbiased look at dataset bias,” in CVPR 2011, pp. 1521–1528, IEEE, 2011 | |
| dc.relation.references | H. Song, M. Kim, D. Park, Y. Shin, and J.-G. Lee, “Learning from noisy labels with deep neural networks: A survey,” IEEE transactions on neural networks and learning systems, vol. 34, no. 11, pp. 8135–8153, 2022 | |
| dc.relation.references | M. Minsky and S. A. Papert, Perceptrons, reissue of the 1988 expanded edition with a new foreword by Léon Bottou: an introduction to computational geometry. MIT press, 2017 | |
| dc.relation.references | J. Zhang, X. Lv, Q. Sun, Q. Zhang, X. Wei, and B. Liu, “Sdresu-net: separable and dilated residual u-net for mri brain tumor segmentation,” Current Medical Imaging, vol. 16, no. 6, pp. 720–728, 2020 | |
| dc.rights.accessrights | info:eu-repo/semantics/openAccess | |
| dc.rights.license | Atribución-NoComercial 4.0 Internacional | |
| dc.rights.uri | http://creativecommons.org/licenses/by-nc/4.0/ | |
| dc.subject.ddc | 000 - Ciencias de la computación, información y obras generales::006 - Métodos especiales de computación | |
| dc.subject.proposal | Artificial inteligent | eng |
| dc.subject.proposal | Deep learning | eng |
| dc.subject.proposal | Computer vision | eng |
| dc.subject.proposal | Mutiple annotators | eng |
| dc.subject.proposal | Truncated generalized cross entropy | eng |
| dc.subject.proposal | Inteligencia artificial | spa |
| dc.subject.proposal | Aprendizaje profundo | spa |
| dc.subject.proposal | Visión por computador | spa |
| dc.subject.proposal | Múltiples anotadores | spa |
| dc.subject.proposal | Entropía cruzada generalizada truncada | spa |
| dc.subject.unesco | Inteligencia artificial | |
| dc.subject.unesco | Artificial intelligence | |
| dc.subject.unesco | Análisis de datos | |
| dc.subject.unesco | Data analysis | |
| dc.title | Regularized lightweight deep learning for semantic image segmentation | eng |
| dc.title.translated | Aprendizaje profundo ligero regularizado para la segmentación semántica de imágenes | spa |
| dc.type | Trabajo de grado - Maestría | |
| dc.type.coar | http://purl.org/coar/resource_type/c_bdcc | |
| dc.type.coarversion | http://purl.org/coar/version/c_ab4af688f83e57aa | |
| dc.type.content | Text | |
| dc.type.driver | info:eu-repo/semantics/masterThesis | |
| dc.type.version | info:eu-repo/semantics/acceptedVersion | |
| dcterms.audience.professionaldevelopment | Bibliotecarios | |
| dcterms.audience.professionaldevelopment | Estudiantes | |
| dcterms.audience.professionaldevelopment | Investigadores | |
| dcterms.audience.professionaldevelopment | Maestros | |
| oaire.accessrights | http://purl.org/coar/access_right/c_abf2 | |
| oaire.awardtitle | Artificial Vision System for Monitoring and Tracking Analgesic and Anesthetic Effects Administered via Neuroaxial Epidural in Obstetric Population during Labor for the Strengthening of Maternal Health Services at Hospital Universitario de Caldas– SES HUC (Hermes 57661) | |
| oaire.fundername | Hospital Universitario de Caldas– SES HUC | |
| oaire.fundername | Universidad Nacional de Colombia sede Manizales | |
| oaire.fundername | Ministerio de Ciencia, Tecnología e Innovación de Colombia |
Archivos
Bloque original
1 - 1 de 1
Cargando...
- Nombre:
- Tesis de Maestría en Ingeniería - Automatización Industrial.pdf
- Tamaño:
- 13.5 MB
- Formato:
- Adobe Portable Document Format
- Descripción:
- Tesis de Maestría en Ingeniería - Automatización Industrial
Bloque de licencias
1 - 1 de 1
Cargando...
- Nombre:
- license.txt
- Tamaño:
- 5.74 KB
- Formato:
- Item-specific license agreed upon to submission
- Descripción:

