This paper introduces a method for automatically designing convolutional neural network (CNN) architectures tailored for diabetic retinopathy (DR) diagnosis using fundus images. The approach uses k-medoid clustering, PCA, and inter/intra-class variations to optimize CNN depth and width. Validated on datasets including a local Saudi dataset and Kaggle benchmarks, the custom-designed models outperform pre-trained CNNs with fewer parameters.
This paper introduces a convolutional transformer model for classifying tomato maturity, along with a new UAE-sourced dataset, KUTomaData, for training segmentation and classification models. The model combines CNNs and transformers and was tested against two public datasets. Results showed state-of-the-art performance, outperforming existing methods by significant margins in mAP scores across all three datasets.
Researchers have developed a CNN-based deep learning model for predicting coastal flooding in cities under various sea-level rise scenarios. The model utilizes a vision-based, low-resource DL framework and is trained on datasets from Abu Dhabi and San Francisco. Results show a 20% reduction in mean absolute error compared to existing methods, demonstrating potential for scalable coastal flood management.
The paper introduces the Prism Hypothesis, which posits a correspondence between an encoder's feature spectrum and its functional role, with semantic encoders capturing low-frequency components and pixel encoders retaining high-frequency information. Based on this, the authors propose Unified Autoencoding (UAE), a model that harmonizes semantic structure and pixel details using a frequency-band modulator. Experiments on ImageNet and MS-COCO demonstrate that UAE effectively unifies semantic abstraction and pixel-level fidelity, achieving state-of-the-art performance.
A new brain tumor segmentation method based on convolutional neural networks is proposed for the BraTS-GoAT challenge. The method employs the MedNeXt architecture and model ensembling to segment tumors in brain MRI scans from diverse populations. Experiments on the unseen validation set demonstrate promising results with an average DSC of 85.54%.
The paper introduces MedNNS, a neural network search framework designed for medical imaging, addressing challenges in architecture selection and weight initialization. MedNNS constructs a meta-space encoding datasets and models based on their performance using a Supernetwork-based approach, expanding the model zoo size by 51x. The framework incorporates rank loss and Fréchet Inception Distance (FID) loss to capture inter-model and inter-dataset relationships, improving alignment in the meta-space and outperforming ImageNet pre-trained DL models and SOTA NAS methods.
The study compares deep learning models trained via transfer learning from ImageNet (TII-models) against those trained solely on medical images (LMI-models) for disease segmentation. Results show that combining outputs from both model types can improve segmentation performance by up to 10% in certain scenarios. A repository of models, code, and over 10,000 medical images is available on GitHub to facilitate further research.
This paper introduces a deep vision-based framework for predicting coastal floods under climate change, addressing the challenges of limited training data and high-dimensional output. The framework employs and compares various deep learning models, including a custom compact CNN architecture, against geostatistical and traditional machine learning methods. A new synthetic dataset of flood inundation maps for Abu Dhabi's coast is also provided to benchmark future models.