Sign in to use this feature.

Years

Between: -

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,289)

Search Parameters:
Journal = J. Imaging

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
Article
Intelligent Performance Evaluation in Rowing Sport Using a Graph-Matching Network
J. Imaging 2023, 9(9), 181; https://doi.org/10.3390/jimaging9090181 - 31 Aug 2023
Abstract
Rowing competitions require consistent rowing strokes among crew members to achieve optimal performance. However, existing motion analysis techniques often rely on wearable sensors, leading to challenges in sporter inconvenience. The aim of our work is to use a graph-matching network to analyze the [...] Read more.
Rowing competitions require consistent rowing strokes among crew members to achieve optimal performance. However, existing motion analysis techniques often rely on wearable sensors, leading to challenges in sporter inconvenience. The aim of our work is to use a graph-matching network to analyze the similarity in rowers’ rowing posture and further pair rowers to improve the performance of their rowing team. This study proposed a novel video-based performance analysis system to analyze paired rowers using a graph-matching network. The proposed system first detected human joint points, as acquired from the OpenPose system, and then the graph embedding model and graph-matching network model were applied to analyze similarities in rowing postures between paired rowers. When analyzing the postures of the paired rowers, the proposed system detected the same starting point of their rowing postures to achieve more accurate pairing results. Finally, variations in the similarities were displayed using the proposed time-period similarity processing. The experimental results show that the proposed time-period similarity processing of the 2D graph-embedding model (GEM) had the best pairing results. Full article
Show Figures

Figure 1

Article
Automatic 3D Postoperative Evaluation of Complex Orthopaedic Interventions
J. Imaging 2023, 9(9), 180; https://doi.org/10.3390/jimaging9090180 - 31 Aug 2023
Abstract
In clinical practice, image-based postoperative evaluation is still performed without state-of-the-art computer methods, as these are not sufficiently automated. In this study we propose a fully automatic 3D postoperative outcome quantification method for the relevant steps of orthopaedic interventions on the example of [...] Read more.
In clinical practice, image-based postoperative evaluation is still performed without state-of-the-art computer methods, as these are not sufficiently automated. In this study we propose a fully automatic 3D postoperative outcome quantification method for the relevant steps of orthopaedic interventions on the example of Periacetabular Osteotomy of Ganz (PAO). A typical orthopaedic intervention involves cutting bone, anatomy manipulation and repositioning as well as implant placement. Our method includes a segmentation based deep learning approach for detection and quantification of the cuts. Furthermore, anatomy repositioning was quantified through a multi-step registration method, which entailed a coarse alignment of the pre- and postoperative CT images followed by a fine fragment alignment of the repositioned anatomy. Implant (i.e., screw) position was identified by 3D Hough transform for line detection combined with fast voxel traversal based on ray tracing. The feasibility of our approach was investigated on 27 interventions and compared against manually performed 3D outcome evaluations. The results show that our method can accurately assess the quality and accuracy of the surgery. Our evaluation of the fragment repositioning showed a cumulative error for the coarse and fine alignment of 2.1 mm. Our evaluation of screw placement accuracy resulted in a distance error of 1.32 mm for screw head location and an angular deviation of 1.1° for screw axis. As a next step we will explore generalisation capabilities by applying the method to different interventions. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Article
Data-Weighted Multivariate Generalized Gaussian Mixture Model: Application to Point Cloud Robust Registration
J. Imaging 2023, 9(9), 179; https://doi.org/10.3390/jimaging9090179 - 31 Aug 2023
Abstract
In this paper, a weighted multivariate generalized Gaussian mixture model combined with stochastic optimization is proposed for point cloud registration. The mixture model parameters of the target scene and the scene to be registered are updated iteratively by the fixed point method under [...] Read more.
In this paper, a weighted multivariate generalized Gaussian mixture model combined with stochastic optimization is proposed for point cloud registration. The mixture model parameters of the target scene and the scene to be registered are updated iteratively by the fixed point method under the framework of the EM algorithm, and the number of components is determined based on the minimum message length criterion (MML). The KL divergence between these two mixture models is utilized as the loss function for stochastic optimization to find the optimal parameters of the transformation model. The self-built point clouds are used to evaluate the performance of the proposed algorithm on rigid registration. Experiments demonstrate that the algorithm dramatically reduces the impact of noise and outliers and effectively extracts the key features of the data-intensive regions. Full article
(This article belongs to the Special Issue Feature Papers in Section AI in Imaging)
Show Figures

Figure 1

Article
2-[18F]FDG-PET/CT in Cancer of Unknown Primary Tumor—A Retrospective Register-Based Cohort Study
J. Imaging 2023, 9(9), 178; https://doi.org/10.3390/jimaging9090178 (registering DOI) - 31 Aug 2023
Viewed by 85
Abstract
We investigated the impact of 2-[18F]FDG-PET/CT on detection rate (DR) of the primary tumor and survival in patients with suspected cancer of unknown primary tumor (CUP), comparing it to the conventional diagnostic imaging method, CT. Patients who received a tentative CUP diagnosis at [...] Read more.
We investigated the impact of 2-[18F]FDG-PET/CT on detection rate (DR) of the primary tumor and survival in patients with suspected cancer of unknown primary tumor (CUP), comparing it to the conventional diagnostic imaging method, CT. Patients who received a tentative CUP diagnosis at Odense University Hospital from 2014–2017 were included. Patients receiving a 2-[18F]FDG-PET/CT were assigned to the 2-[18F]FDG-PET/CT group and patients receiving a CT only to the CT group. DR was calculated as the proportion of true positive findings of 2-[18F]FDG-PET/CT and CT scans, separately, using biopsy of the primary tumor, autopsy, or clinical decision as reference standard. Survival analyses included Kaplan–Meier estimates and Cox proportional hazards regression adjusted for age, sex, treatment, and propensity score. We included 193 patients. Of these, 159 were in the 2-[18F]FDG-PET/CT group and 34 were in the CT group. DR was 36.5% in the 2-[18F]FDG-PET/CT group and 17.6% in the CT group, respectively (p = 0.012). Median survival was 7.4 (95% CI 0.4–98.7) months in the 2-[18F]FDG-PET/CT group and 3.8 (95% CI 0.2–98.1) in the CT group. Survival analysis showed a crude hazard ratio of 0.63 (p = 0.024) and an adjusted hazard ratio of 0.68 (p = 0.087) for the 2-[18F]FDG-PET/CT group compared with CT. This study found a significantly higher DR of the primary tumor in suspected CUP patients using 2-[18F]FDG-PET/CT compared with patients receiving only CT, with possible immense clinical importance. No significant difference in survival was found, although a possible tendency towards longer survival in the 2-[18F]FDG-PET/CT group was observed. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Article
Explainable Artificial Intelligence (XAI) for Deep Learning Based Medical Imaging Classification
J. Imaging 2023, 9(9), 177; https://doi.org/10.3390/jimaging9090177 - 30 Aug 2023
Viewed by 236
Abstract
Recently, deep learning has gained significant attention as a noteworthy division of artificial intelligence (AI) due to its high accuracy and versatile applications. However, one of the major challenges of AI is the need for more interpretability, commonly referred to as the black-box [...] Read more.
Recently, deep learning has gained significant attention as a noteworthy division of artificial intelligence (AI) due to its high accuracy and versatile applications. However, one of the major challenges of AI is the need for more interpretability, commonly referred to as the black-box problem. In this study, we introduce an explainable AI model for medical image classification to enhance the interpretability of the decision-making process. Our approach is based on segmenting the images to provide a better understanding of how the AI model arrives at its results. We evaluated our model on five datasets, including the COVID-19 and Pneumonia Chest X-ray dataset, Chest X-ray (COVID-19 and Pneumonia), COVID-19 Image Dataset (COVID-19, Viral Pneumonia, Normal), and COVID-19 Radiography Database. We achieved testing and validation accuracy of 90.6% on a relatively small dataset of 6432 images. Our proposed model improved accuracy and reduced time complexity, making it more practical for medical diagnosis. Our approach offers a more interpretable and transparent AI model that can enhance the accuracy and efficiency of medical diagnosis. Full article
(This article belongs to the Special Issue Explainable AI for Image-Aided Diagnosis)
Show Figures

Figure 1

Article
Hybrid Autofluorescence and Optoacoustic Microscopy for the Label-Free, Early and Rapid Detection of Pathogenic Infections in Vegetative Tissues
J. Imaging 2023, 9(9), 176; https://doi.org/10.3390/jimaging9090176 - 29 Aug 2023
Viewed by 365
Abstract
Agriculture plays a pivotal role in food security and food security is challenged by pests and pathogens. Due to these challenges, the yields and quality of agricultural production are reduced and, in response, restrictions in the trade of plant products are applied. Governments [...] Read more.
Agriculture plays a pivotal role in food security and food security is challenged by pests and pathogens. Due to these challenges, the yields and quality of agricultural production are reduced and, in response, restrictions in the trade of plant products are applied. Governments have collaborated to establish robust phytosanitary measures, promote disease surveillance, and invest in research and development to mitigate the impact on food security. Classic as well as modernized tools for disease diagnosis and pathogen surveillance do exist, but most of these are time-consuming, laborious, or are less sensitive. To that end, we propose the innovative application of a hybrid imaging approach through the combination of confocal fluorescence and optoacoustic imaging microscopy. This has allowed us to non-destructively detect the physiological changes that occur in plant tissues as a result of a pathogen-induced interaction well before visual symptoms occur. When broccoli leaves were artificially infected with Xanthomonas campestris pv. campestris (Xcc), eventually causing an economically important bacterial disease, the induced optical absorption alterations could be detected at very early stages of infection. Therefore, this innovative microscopy approach was positively utilized to detect the disease caused by a plant pathogen, showing that it can also be employed to detect quarantine pathogens such as Xylella fastidiosa. Full article
(This article belongs to the Special Issue Fluorescence Imaging and Analysis of Cellular System)
Show Figures

Figure 1

Article
End-to-End Depth-Guided Relighting Using Lightweight Deep Learning-Based Method
J. Imaging 2023, 9(9), 175; https://doi.org/10.3390/jimaging9090175 - 28 Aug 2023
Viewed by 167
Abstract
Image relighting, which involves modifying the lighting conditions while preserving the visual content, is fundamental to computer vision. This study introduced a bi-modal lightweight deep learning model for depth-guided relighting. The model utilizes the Res2Net Squeezed block’s ability to capture long-range dependencies and [...] Read more.
Image relighting, which involves modifying the lighting conditions while preserving the visual content, is fundamental to computer vision. This study introduced a bi-modal lightweight deep learning model for depth-guided relighting. The model utilizes the Res2Net Squeezed block’s ability to capture long-range dependencies and to enhance feature representation for both the input image and its corresponding depth map. The proposed model adopts an encoder–decoder structure with Res2Net Squeezed blocks integrated at each stage of encoding and decoding. The model was trained and evaluated on the VIDIT dataset, which consists of 300 triplets of images. Each triplet contains the input image, its corresponding depth map, and the relit image under diverse lighting conditions, such as different illuminant angles and color temperatures. The enhanced feature representation and improved information flow within the Res2Net Squeezed blocks enable the model to handle complex lighting variations and generate realistic relit images. The experimental results demonstrated the proposed approach’s effectiveness in relighting accuracy, measured by metrics such as the PSNR, SSIM, and visual quality. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

Protocol
3D Ultrasound and MRI in Assessing Resection Margins during Tongue Cancer Surgery: A Research Protocol for a Clinical Diagnostic Accuracy Study
J. Imaging 2023, 9(9), 174; https://doi.org/10.3390/jimaging9090174 - 28 Aug 2023
Viewed by 180
Abstract
Surgery is the primary treatment for tongue cancer. The goal is a complete resection of the tumor with an adequate margin of healthy tissue around the tumor.Inadequate margins lead to a high risk of local cancer recurrence and the need for adjuvant therapies. [...] Read more.
Surgery is the primary treatment for tongue cancer. The goal is a complete resection of the tumor with an adequate margin of healthy tissue around the tumor.Inadequate margins lead to a high risk of local cancer recurrence and the need for adjuvant therapies. Ex vivo imaging of the resected surgical specimen has been suggested for margin assessment and improved surgical results. Therefore, we have developed a novel three-dimensional (3D) ultrasound imaging technique to improve the assessment of resection margins during surgery. In this research protocol, we describe a study comparing the accuracy of 3D ultrasound, magnetic resonance imaging (MRI), and clinical examination of the surgical specimen to assess the resection margins during cancer surgery. Tumor segmentation and margin measurement will be performed using 3D ultrasound and MRI of the ex vivo specimen. We will determine the accuracy of each method by comparing the margin measurements and the proportion of correctly classified margins (positive, close, and free) obtained by each technique with respect to the gold standard histopathology. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Article
A Framework for Detecting Thyroid Cancer from Ultrasound and Histopathological Images Using Deep Learning, Meta-Heuristics, and MCDM Algorithms
J. Imaging 2023, 9(9), 173; https://doi.org/10.3390/jimaging9090173 - 27 Aug 2023
Viewed by 282
Abstract
Computer-assisted diagnostic systems have been developed to aid doctors in diagnosing thyroid-related abnormalities. The aim of this research is to improve the diagnosis accuracy of thyroid abnormality detection models that can be utilized to alleviate undue pressure on healthcare professionals. In this research, [...] Read more.
Computer-assisted diagnostic systems have been developed to aid doctors in diagnosing thyroid-related abnormalities. The aim of this research is to improve the diagnosis accuracy of thyroid abnormality detection models that can be utilized to alleviate undue pressure on healthcare professionals. In this research, we proposed deep learning, metaheuristics, and a MCDM algorithms-based framework to detect thyroid-related abnormalities from ultrasound and histopathological images. The proposed method uses three recently developed deep learning techniques (DeiT, Swin Transformer, and Mixer-MLP) to extract features from the thyroid image datasets. The feature extraction techniques are based on the Image Transformer and MLP models. There is a large number of redundant features that can overfit the classifiers and reduce the generalization capabilities of the classifiers. In order to avoid the overfitting problem, six feature transformation techniques (PCA, TSVD, FastICA, ISOMAP, LLE, and UMP) are analyzed to reduce the dimensionality of the data. There are five different classifiers (LR, NB, SVC, KNN, and RF) evaluated using the 5-fold stratified cross-validation technique on the transformed dataset. Both datasets exhibit large class imbalances and hence, the stratified cross-validation technique is used to evaluate the performance. The MEREC-TOPSIS MCDM technique is used for ranking the evaluated models at different analysis stages. In the first stage, the best feature extraction and classification techniques are chosen, whereas, in the second stage, the best dimensionality reduction method is evaluated in wrapper feature selection mode. Two best-ranked models are further selected for the weighted average ensemble learning and features selection using the recently proposed meta-heuristics FOX-optimization algorithm. The PCA+FOX optimization-based feature selection + random forest model achieved the highest TOPSIS score and performed exceptionally well with an accuracy of 99.13%, F2-score of 98.82%, and AUC-ROC score of 99.13% on the ultrasound dataset. Similarly, the model achieved an accuracy score of 90.65%, an F2-score of 92.01%, and an AUC-ROC score of 95.48% on the histopathological dataset. This study exploits the combination novelty of different algorithms in order to improve the thyroid cancer diagnosis capabilities. This proposed framework outperforms the current state-of-the-art diagnostic methods for thyroid-related abnormalities in ultrasound and histopathological datasets and can significantly aid medical professionals by reducing the excessive burden on the medical fraternity. Full article
Show Figures

Figure 1

Article
PP-JPEG: A Privacy-Preserving JPEG Image-Tampering Localization
J. Imaging 2023, 9(9), 172; https://doi.org/10.3390/jimaging9090172 - 27 Aug 2023
Viewed by 266
Abstract
The widespread availability of digital image-processing software has given rise to various forms of image manipulation and forgery, which can pose a significant challenge in different fields, such as law enforcement, journalism, etc. It can also lead to privacy concerns. We are proposing [...] Read more.
The widespread availability of digital image-processing software has given rise to various forms of image manipulation and forgery, which can pose a significant challenge in different fields, such as law enforcement, journalism, etc. It can also lead to privacy concerns. We are proposing that a privacy-preserving framework to encrypt images before processing them is vital to maintain the privacy and confidentiality of sensitive images, especially those used for the purpose of investigation. To address these challenges, we propose a novel solution that detects image forgeries while preserving the privacy of the images. Our method proposes a privacy-preserving framework that encrypts the images before processing them, making it difficult for unauthorized individuals to access them. The proposed method utilizes a compression quality analysis in the encrypted domain to detect the presence of forgeries in images by determining if the forged portion (dummy image) has a compression quality different from that of the original image (featured image) in the encrypted domain. This approach effectively localizes the tampered portions of the image, even for small pixel blocks of size 10×10 in the encrypted domain. Furthermore, the method identifies the featured image’s JPEG quality using the first minima in the energy graph. Full article
(This article belongs to the Topic Computer Vision and Image Processing)
Show Figures

Figure 1

Article
Improving Medical Imaging with Medical Variation Diffusion Model: An Analysis and Evaluation
J. Imaging 2023, 9(9), 171; https://doi.org/10.3390/jimaging9090171 - 25 Aug 2023
Viewed by 294
Abstract
The Medical VDM is an approach for generating medical images that employs variational diffusion models (VDMs) to smooth images while preserving essential features, including edges. The primary goal of the Medical VDM is to enhance the accuracy and reliability of medical image generation. [...] Read more.
The Medical VDM is an approach for generating medical images that employs variational diffusion models (VDMs) to smooth images while preserving essential features, including edges. The primary goal of the Medical VDM is to enhance the accuracy and reliability of medical image generation. In this paper, we present a comprehensive description of the Medical VDM approach and its mathematical foundation, as well as experimental findings that showcase its efficacy in generating high-quality medical images that accurately reflect the underlying anatomy and physiology. Our results reveal that the Medical VDM surpasses current VDM methods in terms of generating faithful medical images, with a reconstruction loss of 0.869, a diffusion loss of 0.0008, and a latent loss of 5.740068×105. Furthermore, we delve into the potential applications of the Medical VDM in clinical settings, such as its utility in medical education and training and its potential to aid clinicians in diagnosis and treatment planning. Additionally, we address the ethical concerns surrounding the use of generated medical images and propose a set of guidelines for their ethical use. By amalgamating the power of VDMs with clinical expertise, our approach constitutes a significant advancement in the field of medical imaging, poised to enhance medical education, research, and clinical practice, ultimately leading to improved patient outcomes. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Project Report
Bayesian Reconstruction Algorithms for Low-Dose Computed Tomography Are Not Yet Suitable in Clinical Context
J. Imaging 2023, 9(9), 170; https://doi.org/10.3390/jimaging9090170 - 23 Aug 2023
Viewed by 384
Abstract
Computed tomography (CT) is a widely used examination technique that usually requires a compromise between image quality and radiation exposure. Reconstruction algorithms aim to reduce radiation exposure while maintaining comparable image quality. Recently, unsupervised deep learning methods have been proposed for this purpose. [...] Read more.
Computed tomography (CT) is a widely used examination technique that usually requires a compromise between image quality and radiation exposure. Reconstruction algorithms aim to reduce radiation exposure while maintaining comparable image quality. Recently, unsupervised deep learning methods have been proposed for this purpose. In this study, a promising sparse-view reconstruction method (posterior temperature optimized Bayesian inverse model; POTOBIM) is tested for its clinical applicability. For this study, 17 whole-body CTs of deceased were performed. In addition to POTOBIM, reconstruction was performed using filtered back projection (FBP). An evaluation was conducted by simulating sinograms and comparing the reconstruction with the original CT slice for each case. A quantitative analysis was performed using peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM). The quality was assessed visually using a modified Ludewig’s scale. In the qualitative evaluation, POTOBIM was rated worse than the reference images in most cases. A partially equivalent image quality could only be achieved with 80 projections per rotation. Quantitatively, POTOBIM does not seem to benefit from more than 60 projections. Although deep learning methods seem suitable to produce better image quality, the investigated algorithm (POTOBIM) is not yet suitable for clinical routine. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Article
An Innovative Faster R-CNN-Based Framework for Breast Cancer Detection in MRI
J. Imaging 2023, 9(9), 169; https://doi.org/10.3390/jimaging9090169 - 23 Aug 2023
Viewed by 254
Abstract
Replacing lung cancer as the most commonly diagnosed cancer globally, breast cancer (BC) today accounts for 1 in 8 cancer diagnoses and a total of 2.3 million new cases in both sexes combined. An estimated 685,000 women died from BC in 2020, corresponding [...] Read more.
Replacing lung cancer as the most commonly diagnosed cancer globally, breast cancer (BC) today accounts for 1 in 8 cancer diagnoses and a total of 2.3 million new cases in both sexes combined. An estimated 685,000 women died from BC in 2020, corresponding to 16% or 1 in every 6 cancer deaths in women. BC represents a quarter of a total of cancer cases in females and by far the most commonly diagnosed cancer in women in 2020. However, when detected in the early stages of the disease, treatment methods have proven to be very effective in increasing life expectancy and, in many cases, patients fully recover. Several medical imaging modalities, such as X-rays Mammography (MG), Ultrasound (US), Computer Tomography (CT), Magnetic Resonance Imaging (MRI), and Digital Tomosynthesis (DT) have been explored to support radiologists/physicians in clinical decision-making workflows for the detection and diagnosis of BC. In this work, we propose a novel Faster R-CNN-based framework to automate the detection of BC pathological Lesions in MRI. As a main contribution, we have developed and experimentally (statistically) validated an innovative method improving the “breast MRI preprocessing phase” to select the patient’s slices (images) and associated bounding boxes representing pathological lesions. In this way, it is possible to create a more robust training (benchmarking) dataset to feed Deep Learning (DL) models, reducing the computation time and the dimension of the dataset, and more importantly, to identify with high accuracy the specific regions (bounding boxes) for each of the patient’s images, in which a possible pathological lesion (tumor) has been identified. As a result, in an experimental setting using a fully annotated dataset (released to the public domain) comprising a total of 922 MRI-based BC patient cases, we have achieved, as the most accurate trained model, an accuracy rate of 97.83%, and subsequently, applying a ten-fold cross-validation method, a mean accuracy on the trained models of 94.46% and an associated standard deviation of 2.43%. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

Article
Multimodal Approach for Enhancing Biometric Authentication
J. Imaging 2023, 9(9), 168; https://doi.org/10.3390/jimaging9090168 - 22 Aug 2023
Viewed by 206
Abstract
Unimodal biometric systems rely on a single source or unique individual biological trait for measurement and examination. Fingerprint-based biometric systems are the most common, but they are vulnerable to presentation attacks or spoofing when a fake fingerprint is presented to the sensor. To [...] Read more.
Unimodal biometric systems rely on a single source or unique individual biological trait for measurement and examination. Fingerprint-based biometric systems are the most common, but they are vulnerable to presentation attacks or spoofing when a fake fingerprint is presented to the sensor. To address this issue, we propose an enhanced biometric system based on a multimodal approach using two types of biological traits. We propose to combine fingerprint and Electrocardiogram (ECG) signals to mitigate spoofing attacks. Specifically, we design a multimodal deep learning architecture that accepts fingerprints and ECG as inputs and fuses the feature vectors using stacking and channel-wise approaches. The feature extraction backbone of the architecture is based on data-efficient transformers. The experimental results demonstrate the promising capabilities of the proposed approach in enhancing the robustness of the system to presentation attacks. Full article
(This article belongs to the Special Issue Multi-Biometric and Multi-Modal Authentication)
Show Figures

Figure 1

Article
Clinical Validation Benchmark Dataset and Expert Performance Baseline for Colorectal Polyp Localization Methods
J. Imaging 2023, 9(9), 167; https://doi.org/10.3390/jimaging9090167 - 22 Aug 2023
Viewed by 337
Abstract
Colorectal cancer is one of the leading death causes worldwide, but, fortunately, early detection highly increases survival rates, with the adenoma detection rate being one surrogate marker for colonoscopy quality. Artificial intelligence and deep learning methods have been applied with great success to [...] Read more.
Colorectal cancer is one of the leading death causes worldwide, but, fortunately, early detection highly increases survival rates, with the adenoma detection rate being one surrogate marker for colonoscopy quality. Artificial intelligence and deep learning methods have been applied with great success to improve polyp detection and localization and, therefore, the adenoma detection rate. In this regard, a comparison with clinical experts is required to prove the added value of the systems. Nevertheless, there is no standardized comparison in a laboratory setting before their clinical validation. The ClinExpPICCOLO comprises 65 unedited endoscopic images that represent the clinical setting. They include white light imaging and narrow band imaging, with one third of the images containing a lesion but, differently to another public datasets, the lesion does not appear well-centered in the image. Together with the dataset, an expert clinical performance baseline has been established with the performance of 146 gastroenterologists, who were required to locate the lesions in the selected images. Results shows statistically significant differences between experience groups. Expert gastroenterologists’ accuracy was 77.74, while sensitivity and specificity were 86.47 and 74.33, respectively. These values can be established as minimum values for a DL method before performing a clinical trial in the hospital setting. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

Back to TopTop