Javascript must be enabled to continue!
Fast VMZ : code enhancements for video mosaicing and summarization
View through CrossRef
Video Mosaicing and Summarization (VMZ) is a novel image processing pipeline that summarizes the content of a long sequence of geospatial or biomedical videos using a few coverage maps or mini mosaics. The existing VMZ algorithm uses Normalized Cross-Correlation (NCC), Structure Tensor (ST), Affine-Invariant SIFT (ASIFT), Speeded up robust features for its feature matching and homography estimation pipeline, which are the most computationally expensive modules in the VMZ pipeline. Due to these long-running compute-intensive modules, the VMZ pipeline is not suitable for real-time mosaic formation in drones or UAVs. For instance, VMZ takes around 4 hours to generate mini-mosaics from an image sequence containing 9291 image frames. The blending algorithms used for mini-mosaic generation suffer from illumination variation due to the illumination difference in image frames. Such illumination inconsistency causes severe problems for biomedical scene understanding where curvilinear or tiny biological structures are present. VMZ pipeline is also dependent on 3rd party libraries not aligned with the flow of VMZ, which introduces redundant computation. One of the main reasons for the slow processing of the VMZ pipeline is not leveraging any parallel processing techniques and available graphics processing hardware. Therefore, the objective of this thesis is mainly three-fold: (i) speeding up the computeintensive and long-running modules in the VMZ pipeline, (ii) modifying the existing libraries and interfaces for better alignment with VMZ workflow, and (iii) resolving the illumination difference problem of the blending algorithms. Selected longrunning modules with the most impact on the overall run-time have been improved using CPU-based Multi-Threading, GPU-based Parallelization, and better integration with the existing VMZ pipeline. An illumination-matched blending algorithm has been proposed to improve the illumination problem. Besides, to evaluate the performance of different blending algorithms, a novel metric named Maximum Overall Illumination Difference (MOID) has been proposed. The improvement of VMZ modules has resulted in more than 100x speed-up in certain modules, with a 4.4x speed-up for the total VMZ run-time. The novel illumination matched blending resulted in a better MOID value for image sequences not having illumination variance in a single frame.
Title: Fast VMZ : code enhancements for video mosaicing and summarization
Description:
Video Mosaicing and Summarization (VMZ) is a novel image processing pipeline that summarizes the content of a long sequence of geospatial or biomedical videos using a few coverage maps or mini mosaics.
The existing VMZ algorithm uses Normalized Cross-Correlation (NCC), Structure Tensor (ST), Affine-Invariant SIFT (ASIFT), Speeded up robust features for its feature matching and homography estimation pipeline, which are the most computationally expensive modules in the VMZ pipeline.
Due to these long-running compute-intensive modules, the VMZ pipeline is not suitable for real-time mosaic formation in drones or UAVs.
For instance, VMZ takes around 4 hours to generate mini-mosaics from an image sequence containing 9291 image frames.
The blending algorithms used for mini-mosaic generation suffer from illumination variation due to the illumination difference in image frames.
Such illumination inconsistency causes severe problems for biomedical scene understanding where curvilinear or tiny biological structures are present.
VMZ pipeline is also dependent on 3rd party libraries not aligned with the flow of VMZ, which introduces redundant computation.
One of the main reasons for the slow processing of the VMZ pipeline is not leveraging any parallel processing techniques and available graphics processing hardware.
Therefore, the objective of this thesis is mainly three-fold: (i) speeding up the computeintensive and long-running modules in the VMZ pipeline, (ii) modifying the existing libraries and interfaces for better alignment with VMZ workflow, and (iii) resolving the illumination difference problem of the blending algorithms.
Selected longrunning modules with the most impact on the overall run-time have been improved using CPU-based Multi-Threading, GPU-based Parallelization, and better integration with the existing VMZ pipeline.
An illumination-matched blending algorithm has been proposed to improve the illumination problem.
Besides, to evaluate the performance of different blending algorithms, a novel metric named Maximum Overall Illumination Difference (MOID) has been proposed.
The improvement of VMZ modules has resulted in more than 100x speed-up in certain modules, with a 4.
4x speed-up for the total VMZ run-time.
The novel illumination matched blending resulted in a better MOID value for image sequences not having illumination variance in a single frame.
Related Results
Joint Beamforming and Aerial IRS Positioning Design for IRS-assisted MISO System with Multiple Access Points
Joint Beamforming and Aerial IRS Positioning Design for IRS-assisted MISO System with Multiple Access Points
<p><code>Intelligent reflecting surface (IRS) is a promising concept for </code><code><u>6G</u></code><code> wireless communications...
Joint Beamforming and Aerial IRS Positioning Design for IRS-assisted MISO System with Multiple Access Points
Joint Beamforming and Aerial IRS Positioning Design for IRS-assisted MISO System with Multiple Access Points
<p><code>Intelligent reflecting surface (IRS) is a promising concept for </code><code><u>6G</u></code><code> wireless communications...
Sentencing Enhancements
Sentencing Enhancements
Sentencing enhancements are policies that mandate that people who are convicted of criminalized behaviors while engaging in generally non-criminalized behaviors—such as being in a ...
CMFF_VS:A Video Summarization Extraction Model based on Cross-modal Feature Fusion
CMFF_VS:A Video Summarization Extraction Model based on Cross-modal Feature Fusion
Abstract
Video summarization aims to present the most relevant and important information in the video stream in the form of a summary. Most existing researches focus on the...
Exploring Summarization Performance: A Comparison of Pointer Generator, Pegasus, and GPT-3 Models
Exploring Summarization Performance: A Comparison of Pointer Generator, Pegasus, and GPT-3 Models
The world is rapidly advancing technologically and the way we communicate is changing with it.We are now able to send messages through text, voice, or video chat, which means that ...
Divinator: A Visual Studio Code Extension to Source Code Summarization
Divinator: A Visual Studio Code Extension to Source Code Summarization
Software developers spend a substantial amount of time reading and understanding code. Research has shown that code comprehension tasks can be expedited by reading the available do...
NETWORK VIDEO CONTENT AS A FORM OF UNIVERSITY PROMOTION
NETWORK VIDEO CONTENT AS A FORM OF UNIVERSITY PROMOTION
In the context of visualization and digitalization of media consumption, network video content is becoming an important form of university promotion in the educational services mar...
Enhancing Real-Time Video Processing With Artificial Intelligence: Overcoming Resolution Loss, Motion Artifacts, And Temporal Inconsistencies
Enhancing Real-Time Video Processing With Artificial Intelligence: Overcoming Resolution Loss, Motion Artifacts, And Temporal Inconsistencies
Purpose: Traditional video processing techniques often struggle with critical challenges such as low resolution, motion artifacts, and temporal inconsistencies, especially in real-...

