

de Recherche et d’Innovation
en Cybersécurité et Société
Amirkhani, D.; Allili, M. S.; Lapointe, J. -F.
CrackSight: An Efficient Crack Segmentation Model in Varying Acquisition Ranges and Complex Backgrounds Article de journal
Dans: IEEE Transactions on Automation Science and Engineering, vol. 22, p. 19197–19214, 2025, ISSN: 15455955 (ISSN).
Résumé | Liens | BibTeX | Étiquettes: Attention mechanisms, Codes (symbols), Complex background, complex backgrounds, Crack detection, Crack propagation, Crack segmentation, Crack segmentations, Detection features, End to end, Feature extraction, Features extraction, Global context, Image segmentation, Learning models, Learning systems, Segmentation models, Transportation infrastructures
@article{amirkhani_cracksight_2025,
title = {CrackSight: An Efficient Crack Segmentation Model in Varying Acquisition Ranges and Complex Backgrounds},
author = {D. Amirkhani and M. S. Allili and J. -F. Lapointe},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-105011756992&doi=10.1109%2FTASE.2025.3591407&partnerID=40&md5=d908b79e863a4725d10bec325b761f34},
doi = {10.1109/TASE.2025.3591407},
issn = {15455955 (ISSN)},
year = {2025},
date = {2025-01-01},
journal = {IEEE Transactions on Automation Science and Engineering},
volume = {22},
pages = {19197–19214},
abstract = {Accurate crack segmentation in concrete transportation infrastructures is critical for ensuring structural integrity and facilitating timely maintenance interventions. This paper presents CrackSight, an end-to-end deep learning model for precise crack segmentation across varying observational ranges and extremely complex backgrounds. CrackSight seamlessly integrates crack detection and segmentation through two branches. The Detection Feature Extraction Branch (DFEB) provides global context for crack localization in complex backgrounds or at far observation ranges. It guides the segmentation model to focus on regions with the highest crack-prone potential. The segmentation branch leverages the fusion of multi-scale feature maps using dilated convolutions, allowing to capture subtle and complex crack patterns. The branch also incorporates the Dual-Attention Linear Focus Mechanism (DALFM) enhancing crack segmentation through saliency-driven improvements. Finally, CrackSight uses a novel hybrid contextual loss, which dynamically compensates for class imbalance and enhance crack discrimination against complex backgrounds. Our model is also lightweight and can be run in resource-constrained environments, making it suitable for real-world inspection using mobile platforms. Our results demonstrate that it significantly improves segmentation accuracy, setting a new benchmark for crack segmentation. The dataset and additional resources are available on GitHub. Note to Practitioners—CrackSight is a dual-branch deep learning framework designed for accurate and efficient segmentation of concrete cracks under challenging real-world conditions. By combining a detection-guided localization branch with a context-aware segmentation, CrackSight offers enhanced robustness to noise, background clutter, and varying acquisition distances, common challenges in UAV-based infrastructure inspections. Its architecture integrates multi-scale feature fusion and adaptive contextual guidance, enabling reliable detection of both fine and fragmented cracks. With its lightweight design and fast inference time, CrackSight offers practitioners a practical and scalable solution for automating visual inspection tasks, reducing manual effort, and improving safety in structural health monitoring workflows. © 2025 IEEE.},
keywords = {Attention mechanisms, Codes (symbols), Complex background, complex backgrounds, Crack detection, Crack propagation, Crack segmentation, Crack segmentations, Detection features, End to end, Feature extraction, Features extraction, Global context, Image segmentation, Learning models, Learning systems, Segmentation models, Transportation infrastructures},
pubstate = {published},
tppubtype = {article}
}
Allaoui, M. L.; Allili, M. S.
MixLVMM: A Mixture of Lightweight Vision Mamba Model for Enhancing Skin Lesion Segmentation Across High Tone Variability Article de journal
Dans: IEEE Access, vol. 13, p. 121234–121249, 2025, ISSN: 21693536 (ISSN).
Résumé | Liens | BibTeX | Étiquettes: Attention mechanism, Attention mechanisms, Computational efficiency, Critical challenges, Dermatology, Diagnosis, Image segmentation, Lesion segmentations, Lung cancer, Mixture of experts model, Mixture-of-experts model, Segmentation performance, Skin lesion, Skin lesion segmentation, Skin/lesion tone variability, Vision mamba
@article{allaoui_mixlvmm_2025,
title = {MixLVMM: A Mixture of Lightweight Vision Mamba Model for Enhancing Skin Lesion Segmentation Across High Tone Variability},
author = {M. L. Allaoui and M. S. Allili},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-105012036322&doi=10.1109%2FACCESS.2025.3588476&partnerID=40&md5=1cf51dcf43653e1677ad36a1360392ac},
doi = {10.1109/ACCESS.2025.3588476},
issn = {21693536 (ISSN)},
year = {2025},
date = {2025-01-01},
journal = {IEEE Access},
volume = {13},
pages = {121234–121249},
abstract = {Accurate skin lesion segmentation remains a critical challenge in automated dermatological diagnosis due to heterogeneous lesion presentations, ambiguous boundaries, imaging artifacts, and significant variability in skin and lesion tones across diverse populations. Current segmentation methods inadequately address these multifaceted complexities, particularly failing to handle extreme tone variations that can lead to diagnostic bias. To address these limitations, we present the Mixture of Lightweight Vision Mamba Model (MixLVMM), a novel expert-based framework that enhances segmentation performance across high tone variability through specialized processing. Our approach employs a Siamese network with triplet loss as a gate mechanism to categorize lesions based on tonal characteristics, routing each image to specialized Vision Mamba Model (VMM) experts optimized for specific lesion categories. Each expert utilizes a U-shaped architecture incorporating Focused Vision Mamba blocks and Adaptive Salient Region Attention modules to capture lesion-specific features while maintaining computational efficiency. Comprehensive evaluation on ISIC and PH2 datasets demonstrates that MixLVMM achieves superior segmentation performance with an average Dice coefficient of 93%, surpassing state-of-the-art methods while maintaining efficiency with only 2.5M parameters. These results establish MixLVMM as a robust solution for addressing tone-related segmentation challenges in clinical dermatology, offering both high accuracy and practical deployment feasibility for real-world applications. © 2013 IEEE.},
keywords = {Attention mechanism, Attention mechanisms, Computational efficiency, Critical challenges, Dermatology, Diagnosis, Image segmentation, Lesion segmentations, Lung cancer, Mixture of experts model, Mixture-of-experts model, Segmentation performance, Skin lesion, Skin lesion segmentation, Skin/lesion tone variability, Vision mamba},
pubstate = {published},
tppubtype = {article}
}
Allaoui, M. L.; Allili, M. S.
MEDiXNet: A Robust Mixture of Expert Dermatological Imaging Networks for Skin Lesion Segmentation Article d'actes
Dans: IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recogn., IEEE Computer Society, 2024, ISBN: 19457928 (ISSN); 979-835031333-8 (ISBN), (Journal Abbreviation: IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recogn.).
Résumé | Liens | BibTeX | Étiquettes: Attention mechanism, Attention mechanisms, Blurred boundaries, Cancer detection, Deep learning, Dermatology, Expert systems, Image segmentation, Lesion segmentations, Mixture of experts, Mixture of experts model, Mixture-of-experts model, Salient regions, Skin cancers, Skin lesion, Skin lesion segmentation
@inproceedings{allaoui_medixnet_2024,
title = {MEDiXNet: A Robust Mixture of Expert Dermatological Imaging Networks for Skin Lesion Segmentation},
author = {M. L. Allaoui and M. S. Allili},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-85203397643&doi=10.1109%2fISBI56570.2024.10635430&partnerID=40&md5=c95dd2122f03c944e945b684a111e741},
doi = {10.1109/ISBI56570.2024.10635430},
isbn = {19457928 (ISSN); 979-835031333-8 (ISBN)},
year = {2024},
date = {2024-01-01},
booktitle = {IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recogn.},
publisher = {IEEE Computer Society},
abstract = {Skin lesion segmentation in dermatological imaging is crucial for early skin cancer detection. However, it is challenging due to variation in lesion appearance, blurred boundaries, and the presence of artifacts. Existing segmentation methods often fall short in accurately addressing these issues. We present MEDiXNet, a novel deep learning model combining expert networks with the Adaptive Salient Region Attention Module (ASRAM) to specifically tackle these challenges. Tailored for varying lesion types, MEDiXNet leverages ASRAM to enhance focus on critical regions, substantially improving segmentation accuracy. Tested on the ISIC datasets, it achieved a 94% Dice coefficient, surpassing state-of-the-art methods. MEDiXNet's innovative approach represents a significant advancement in dermatological imaging, promising to elevate the precision of skin cancer diagnostics. © 2024 IEEE.},
note = {Journal Abbreviation: IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recogn.},
keywords = {Attention mechanism, Attention mechanisms, Blurred boundaries, Cancer detection, Deep learning, Dermatology, Expert systems, Image segmentation, Lesion segmentations, Mixture of experts, Mixture of experts model, Mixture-of-experts model, Salient regions, Skin cancers, Skin lesion, Skin lesion segmentation},
pubstate = {published},
tppubtype = {inproceedings}
}



