original video
Recently Published Documents


TOTAL DOCUMENTS

87
(FIVE YEARS 37)

H-INDEX

8
(FIVE YEARS 1)

Author(s):  
D. Minola Davids ◽  
C. Seldev Christopher

The visual data attained from surveillance single-camera or multi-view camera networks is exponentially increasing every day. Identifying the important shots in the presented video which faithfully signify the original video is the major task in video summarization. For executing efficient video summarization of the surveillance systems, optimization algorithm like LFOB-COA is proposed in this paper. Data collection, pre-processing, deep feature extraction (FE), shot segmentation JSFCM, classification using Rectified Linear Unit activated BLSTM, and LFOB-COA are the proposed method’s five steps. Finally a post-processing step is utilized. For recognizing the proposed method’s effectiveness, the results are then contrasted with the existent methods.


Electronics ◽  
2021 ◽  
Vol 10 (22) ◽  
pp. 2768
Author(s):  
Domonkos Varga

No-reference video quality assessment (NR-VQA) has piqued the scientific community’s interest throughout the last few decades, owing to its importance in human-centered interfaces. The goal of NR-VQA is to predict the perceptual quality of digital videos without any information about their distortion-free counterparts. Over the past few decades, NR-VQA has become a very popular research topic due to the spread of multimedia content and video databases. For successful video quality evaluation, creating an effective video representation from the original video is a crucial step. In this paper, we propose a powerful feature vector for NR-VQA inspired by Benford’s law. Specifically, it is demonstrated that first-digit distributions extracted from different transform domains of the video volume data are quality-aware features and can be effectively mapped onto perceptual quality scores. Extensive experiments were carried out on two large, authentically distorted VQA benchmark databases.


2021 ◽  
Author(s):  
Zhongze Lv ◽  
Hu Guan ◽  
Ying Huang ◽  
Shuwu Zhang
Keyword(s):  

2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Hui Qian ◽  
Mengxuan Dai ◽  
Yong Ma ◽  
Jiale Zhao ◽  
Qinghua Liu ◽  
...  

Video situational information detection is widely used in the fields of video query, character anomaly detection, surveillance analysis, and so on. However, most of the existing researches pay much attention to the subject or video backgrounds, but little attention to the recognition of situational information. What is more, because there is no strong relation between the pixel information and the scene information of video data, it is difficult for computers to obtain corresponding high-level scene information through the low-level pixel information of video data. Video scene information detection is mainly to detect and analyze the multiple features in the video and mark the scenes in the video. It is aimed at automatically extracting video scene information from all kinds of original video data and realizing the recognition of scene information through “comprehensive consideration of pixel information and spatiotemporal continuity.” In order to solve the problem of transforming pixel information into scene information, this paper proposes a video scene information detection method based on entity recognition. This model integrates the spatiotemporal relationship between the video subject and object on the basis of entity recognition, so as to realize the recognition of scene information by establishing mapping relation. The effectiveness and accuracy of the model are verified by simulation experiments with the TV series as experimental data. The accuracy of this model in the simulation experiment can reach more than 85%.


2021 ◽  
pp. 648-654
Author(s):  
Rubel Biswas ◽  
Deisy Chaves ◽  
Laura Fernández-Robles ◽  
Eduardo Fidalgo ◽  
Enrique Alegre

Identifying key content from a video is essential for many security applications such as motion/action detection, person re-identification and recognition. Moreover, summarizing the key information from Child Sexual Exploitation Materials, especially videos, which mainly contain distinctive scenes including people’s faces is crucial to speed-up the investigation of Law Enforcement Agencies. In this paper, we present a video summarization strategy that combines perceptual hashing and face detection algorithms to keep the most relevant frames of a video containing people’s faces that may correspond to victims or offenders. Due to legal constraints to access Child Sexual Abuse datasets, we evaluated the performance of the proposed strategy during the detection of adult pornography content with the NDPI-800 dataset. Also, we assessed the capability of our strategy to create video summaries preserving frames with distinctive faces from the original video using ten additional short videos manually labeled. Results showed that our approach can detect pornography content with an accuracy of 84.15% at a speed of 8.05 ms/frame making this appropriate for realtime applications.


2021 ◽  
Vol 3 (9(111)) ◽  
pp. 103-115
Author(s):  
Vladimir Barannik ◽  
Serhii Sidchenko ◽  
Natalia Barannik ◽  
Valeriy Barannik

The demand for image confidentiality is constantly growing. At the same time, ensuring the confidentiality of video information must be organized subject to ensuring its reliability with a given time delay in processing and transmission. Methods of cryptocompression representation of images can be used to solve this problem. They are designed to simultaneously provide compression and protection of video information. The service component is used as the key of the cryptocompression transformation. However, it has a significant volume. It is 25 % of the original video data volume. A method for coding systems of service components in a differentiated basis on the second cascade of cryptocompression representation of images has been developed. The method is based on the developed scheme of data linearization from three-dimensional coordinates of representation in a two-dimensional matrix into a one-dimensional coordinate for one-to-one representation of this element in a vector. Linearization is organized horizontally line by line. On the basis of the developed method, a non-deterministic number of code values of information components is formed. They have non-deterministic lengths and are formed on a non-deterministic number of elements. The uncertainty of positioning of cryptocompression codograms in the general code stream is provided, which virtually eliminates the possibility of their unauthorized decryption. The method provides a reduction in the volume of the service component of the cryptocompression codogram. The service data volume is 6.25 % of the original video data volume. The method provides an additional reduction in the volume of cryptocompression representation of images without loss of information quality relative to the original video data on average from 1.08 to 1.54 times, depending on the degree of their saturation


2021 ◽  
pp. 1-20
Author(s):  
Himani Sharma ◽  
Navdeep Kanwal

Multimedia communication as well as other related innovations are gaining tremendous growth in the modern technological era. Even though digital content has traditionally proved to be a piece of legitimate evidence. But the latest technologies have lessened this trust, as a variety of video editing tools have been developed to modify the original video. Therefore, in order to resolve this problem, a new technique has been proposed for the detection of duplicate video sequences. The present paper utilizes gray values to extract Hu moment features in the current frame. These features are further used for classification of video as authentic or forged. Afterwards there was also need to validate the proposed technique using training and test dataset. But the scarcity of training and test datasets, however, is indeed one of the key problems to validate the effectiveness of video tampering detection techniques. In this perspective, the Video Forensics Library for Frame Duplication (VLFD) dataset has been introduced for frame duplication detection purposes. The proposed dataset is made of 210 native videos, in Ultra-HD and Full-HD resolution, captured with different cameras. Every video is 6 to 15 seconds in length and runs at 30 frames per second. All the recordings have been acquired in three different scenarios (indoor, outdoor, nature) and in landscape mode(s). VLFD includes both authentic and manipulated video files. This dataset has been created as an initial repository for manipulated video and enhanced with new features and new techniques in future.


2021 ◽  
Vol 11 (11) ◽  
pp. 5260
Author(s):  
Theodoros Psallidas ◽  
Panagiotis Koromilas ◽  
Theodoros Giannakopoulos ◽  
Evaggelos Spyrou

The exponential growth of user-generated content has increased the need for efficient video summarization schemes. However, most approaches underestimate the power of aural features, while they are designed to work mainly on commercial/professional videos. In this work, we present an approach that uses both aural and visual features in order to create video summaries from user-generated videos. Our approach produces dynamic video summaries, that is, comprising the most “important” parts of the original video, which are arranged so as to preserve their temporal order. We use supervised knowledge from both the aforementioned modalities and train a binary classifier, which learns to recognize the important parts of videos. Moreover, we present a novel user-generated dataset which contains videos from several categories. Every 1 sec part of each video from our dataset has been annotated by more than three annotators as being important or not. We evaluate our approach using several classification strategies based on audio, video and fused features. Our experimental results illustrate the potential of our approach.


Materials ◽  
2021 ◽  
Vol 14 (10) ◽  
pp. 2575
Author(s):  
Hao Wen ◽  
Chang Huang ◽  
Shengmin Guo

Cracks and pores are two common defects in metallic additive manufacturing (AM) parts. In this paper, deep learning-based image analysis is performed for defect (cracks and pores) classification/detection based on SEM images of metallic AM parts. Three different levels of complexities, namely, defect classification, defect detection and defect image segmentation, are successfully achieved using a simple CNN model, the YOLOv4 model and the Detectron2 object detection library, respectively. The tuned CNN model can classify any single defect as either a crack or pore at almost 100% accuracy. The other two models can identify more than 90% of the cracks and pores in the testing images. In addition to the application of static image analysis, defect detection is also successfully applied on a video which mimics the AM process control images. The trained Detectron2 model can identify almost all the pores and cracks that exist in the original video. This study lays a foundation for future in situ process monitoring of the 3D printing process.


2021 ◽  
Vol 5 (2) ◽  
pp. 400-406
Author(s):  
Alfiansyah Imanda Putra Alfian ◽  
Rusydi Umar ◽  
Abdul Fadlil

The development of digital video technology which is increasingly advanced makes digital video engineering crimes prone to occur. The change in digital video has changed information communication, and it is easy to use in digital crime. One way to solve this digital crime case is to use the NIST (National Institute of Standards and Technology) method for video forensics. The initial stage is carried out by collecting data and carrying out the process of extracting the collected results. A local hash and noise algorithm can then be used to analyze the resulting results, which will detect any digital video interference or manipulation at each video frame, and perform hash analysis to detect the authenticity of the video. In digital video engineering, histogram analysis can be performed by calculating the histogram value metric, which is used to compare the histogram values ​​of the original video and video noise and make graphical comparisons. The results of the difference in frame analysis show that the results of the video show that the 2nd to 7th frames experience an attack while the histogram calculation of the original video centroid value and video tampering results in different values ​​in the third frame, namely with a value of 124.318 and the 7th frame of the video experiencing a difference in the value of 105,966 videos. tampering and 107,456 in the original video. Hash analysis on video tampering results in an invalid SHA-1 hash, this can prove that the video has been manipulated.


Sign in / Sign up

Export Citation Format

Share Document