Search engine for discovering works of Art, research articles, and books related to Art and Culture
ShareThis
Javascript must be enabled to continue!

SelectStitch: Automated Frame Segmentation and Stitching to Create Composite Images from Otoscope Video Clips

View through CrossRef
Background and Objective: the aim of this study is to develop and validate an automated image segmentation-based frame selection and stitching framework to create enhanced composite images from otoscope videos. The proposed framework, called SelectStitch, is useful for classifying eardrum abnormalities using a single composite image instead of the entire raw otoscope video dataset. Methods: SelectStitch consists of a convolutional neural network (CNN) based semantic segmentation approach to detect the eardrum in each frame of the otoscope video, and a stitching engine to generate a high-quality composite image from the detected eardrum regions. In this study, we utilize two separate datasets: the first one has 36 otoscope videos that were used to train a semantic segmentation model, and the second one, containing 100 videos, which was used to test the proposed method. Cases from both adult and pediatric patients were used in this study. A configuration of 4-levels depth U-Net architecture was trained to automatically find eardrum regions in each otoscope video frame from the first dataset. After the segmentation, we automatically selected meaningful frames from otoscope videos by using a pre-defined threshold, i.e., it should contain at least an eardrum region of 20% of a frame size. We have generated 100 composite images from the test dataset. Three ear, nose, and throat (ENT) specialists (ENT-I, ENT-II, ENT-III) compared in two rounds the composite images produced by SelectStitch against the composite images that were generated by the base processes, i.e., stitching all the frames from the same video data, in terms of their diagnostic capabilities. Results: In the first round of the study, ENT-I, ENT-II, ENT-III graded improvement for 58, 57, and 71 composite images out of 100, respectively, for SelectStitch over the base composite, reflecting greater diagnostic capabilities. In the repeat assessment, these numbers were 56, 56, and 64, respectively. We observed that only 6%, 3%, and 3% of the cases received a lesser score than the base composite images, respectively, for ENT-I, ENT-II, and ENT-III in Round-1, and 4%, 0%, and 2% of the cases in Round-2. Conclusions: We conclude that the frame selection and stitching will increase the probability of detecting a lesion even if it appears in a few frames.
Title: SelectStitch: Automated Frame Segmentation and Stitching to Create Composite Images from Otoscope Video Clips
Description:
Background and Objective: the aim of this study is to develop and validate an automated image segmentation-based frame selection and stitching framework to create enhanced composite images from otoscope videos.
The proposed framework, called SelectStitch, is useful for classifying eardrum abnormalities using a single composite image instead of the entire raw otoscope video dataset.
Methods: SelectStitch consists of a convolutional neural network (CNN) based semantic segmentation approach to detect the eardrum in each frame of the otoscope video, and a stitching engine to generate a high-quality composite image from the detected eardrum regions.
In this study, we utilize two separate datasets: the first one has 36 otoscope videos that were used to train a semantic segmentation model, and the second one, containing 100 videos, which was used to test the proposed method.
Cases from both adult and pediatric patients were used in this study.
A configuration of 4-levels depth U-Net architecture was trained to automatically find eardrum regions in each otoscope video frame from the first dataset.
After the segmentation, we automatically selected meaningful frames from otoscope videos by using a pre-defined threshold, i.
e.
, it should contain at least an eardrum region of 20% of a frame size.
We have generated 100 composite images from the test dataset.
Three ear, nose, and throat (ENT) specialists (ENT-I, ENT-II, ENT-III) compared in two rounds the composite images produced by SelectStitch against the composite images that were generated by the base processes, i.
e.
, stitching all the frames from the same video data, in terms of their diagnostic capabilities.
Results: In the first round of the study, ENT-I, ENT-II, ENT-III graded improvement for 58, 57, and 71 composite images out of 100, respectively, for SelectStitch over the base composite, reflecting greater diagnostic capabilities.
In the repeat assessment, these numbers were 56, 56, and 64, respectively.
We observed that only 6%, 3%, and 3% of the cases received a lesser score than the base composite images, respectively, for ENT-I, ENT-II, and ENT-III in Round-1, and 4%, 0%, and 2% of the cases in Round-2.
Conclusions: We conclude that the frame selection and stitching will increase the probability of detecting a lesion even if it appears in a few frames.

Related Results

SelectStitch: Automated Frame Segmentation and Stitching to Create Composite Images from Otoscope Video Clips
SelectStitch: Automated Frame Segmentation and Stitching to Create Composite Images from Otoscope Video Clips
AbstractBackground and ObjectiveThe aim of this study is to develop and validate an automated image segmentation-based frame selection and stitching framework to create enhanced co...
Detection method of PCB component based on automatic optical stitching algorithm
Detection method of PCB component based on automatic optical stitching algorithm
Purpose – This paper aims to propose image stitching by reduction of full line and taking line image as registration image to solve the problem of automatic optic i...
Precision health and deafness–optimizing genetic diagnosis
Precision health and deafness–optimizing genetic diagnosis
<p>Deafness is the most common sensory deficit in humans. In the United States 1-2 in a thousand babies is born with significant deafness, well over half of which is heredita...
Polymer versus Titanium Clips in Laparoscopic Cholecystectomy
Polymer versus Titanium Clips in Laparoscopic Cholecystectomy
Background: Surgical clips are commonly used during laparoscopic cholecystectomy for cystic duct and artery ligation. Titanium and polymer clips are the two most common types used ...
Image and video object segmentation in low supervision scenarios
Image and video object segmentation in low supervision scenarios
Computer vision plays a key role in Artificial Intelligence because of the rich semantic information contained in pixels and the ubiquity of cameras nowadays. Multimedia content is...
Multiple surface segmentation using novel deep learning and graph based methods
Multiple surface segmentation using novel deep learning and graph based methods
<p>The task of automatically segmenting 3-D surfaces representing object boundaries is important in quantitative analysis of volumetric images, which plays a vital role in nu...
AI‐enabled precise brain tumor segmentation by integrating Refinenet and contour‐constrained features in MRI images
AI‐enabled precise brain tumor segmentation by integrating Refinenet and contour‐constrained features in MRI images
AbstractBackgroundMedical image segmentation is a fundamental task in medical image analysis and has been widely applied in multiple medical fields. The latest transformer‐based de...
Time-based Calibration: A Way to Ensure that Stitched Images are Captured Simultaneously
Time-based Calibration: A Way to Ensure that Stitched Images are Captured Simultaneously
<p>With the rapid development of modern science and technology, people&rsquo;s demand for such information as images and videos is also growing, and the requirements for ...

Back to Top