Block-Based Motion Estimation

Author(s):  
Shaifali Madan Arora ◽  
Kavita Khanna

Recent years have witnessed a great technological evolution in video display and capturing technologies leading to the development of new standards of video coding including MPEG-X, H.26X and HEVC. The cost of computations, storage and high bandwidth requirements makes a video data expensive in terms of transmission and storage. This makes video compression absolutely necessary prior to its transmission in order to accommodate for different transmission media's capabilities. Digital video compression technologies therefore have become an important part of the way we create, present, communicate and use visual information. The main aim behind a video compression system is to eliminate the redundancies from a raw video signal. The tradeoff involved in the process of video compression is between the speed, quality and resource utilization. The current chapter explores the techniques, challenges, issues and problems in video compression in detail along with the major advancements in the field.

Author(s):  
Shaifali Madan Arora ◽  
Kavita Khanna

Recent years have witnessed a great technological evolution in video display and capturing technologies leading to the development of new standards of video coding including MPEG-X, H.26X and HEVC. The cost of computations, storage and high bandwidth requirements makes a video data expensive in terms of transmission and storage. This makes video compression absolutely necessary prior to its transmission in order to accommodate for different transmission media's capabilities. Digital video compression technologies therefore have become an important part of the way we create, present, communicate and use visual information. The main aim behind a video compression system is to eliminate the redundancies from a raw video signal. The tradeoff involved in the process of video compression is between the speed, quality and resource utilization. The current chapter explores the techniques, challenges, issues and problems in video compression in detail along with the major advancements in the field.


Author(s):  
O. Gertsiy

The main characteristics of graphic information compression methods with losses and without losses (RLE, LZW, Huffman's method, DEFLATE, JBIG, JPEG, JPEG 2000, Lossless JPEG, fractal and Wawelet) are analyzed in the article. Effective transmission and storage of images in railway communication systems is an important task now. Because large images require large storage resources. This task has become very important in recent years, as the problems of information transmission by telecommunication channels of the transport infrastructure have become urgent. There is also a great need for video conferencing, where the task is to effectively compress video data - because the greater the amount of data, the greater the cost of transmitting information, respectively. Therefore, the use of image compression methods that reduce the file size is the solution to this task. The study highlights the advantages and disadvantages of compression methods. The comparative analysis the basic possibilities of compression methods of graphic information is carried out. The relevance lies in the efficient transfer and storage of graphical information, as big data requires large resources for storage. The practical significance lies in solving the problem of effectively reducing the data size by applying known compression methods.


Author(s):  
V.V.. S.V.S RAMACHANDRAM ◽  
DANIEL N. FINNEY

Video compression is necessary in a wide range of applications to reduce the total data amount required for transmitting or storing video data. Among the coding systems, Motion Estimation is of priority concern in exploiting the temporal redundancy between successive frames, yet also the most time consuming aspect of coding. This paper presents an error detection and data recovery (EDDR) design, based on the residue-and quotient (RQ) code that is embed into ME for video coding testing applications. Based on the Concurrent Error Detection (CED) concept, this work develops a robust EDDR architecture based on the RQ code to detect errors and recovery data in PEs of a ME and, in doing so, further guarantee the excellent reliability for video coding applications. We synthesized this design using Xilinx tool.


Author(s):  
Le Dao Thi Hue ◽  
Luong Pham Van ◽  
Duong Dinh Trieu ◽  
Xiem HoangVan

Video surveillance has been playing an important role in public safety and privacy protection in recent years thanks to its capability of providing the activity monitoring and content analyzing. However, the data associated with long hours surveillance video is huge, making it less attractive to practical applications. In this paper, we propose a low complexity, yet efficient scalable video coding solution for video surveillance system. The proposed surveillance video compression scheme is able to provide the quality scalability feature by following a layered coding structure that consists of one or several enhancement layers on the top of a base layer. In addition, to maintain the backward compatibility with the current video coding standards, the state-of-the-art video coding standard, i.e., High Efficiency Video Coding (HEVC), is employed in the proposed coding solution to compress the base layer. To satisfy the low complexity requirement of the encoder for the video surveillance systems, the distributed coding concept is employed at the enhancement layers. Experiments conducted for a rich set of surveillance video data shown that the proposed surveillance - distributed scalable video coding (S-DSVC) solution significantly outperforms relevant video coding benchmarks, notably the SHVC standard and the HEVC-simulcasting while requiring much lower computational complexity at the encoder which is essential for practical video surveillance applications.


2018 ◽  
Vol 7 (4.10) ◽  
pp. 928
Author(s):  
Prayline Rajabai C ◽  
Sivanantham S

Various video coding standards like H.264 and H.265 are used for video compression and decompression. These coding standards use multiple modules to perform video compression. Motion Estimation (ME) is one of the critical blocks in the video codec which requires extensive computation. Hence it is computationally complex, it critically consumes a massive amount of time to process the video data. Motion Estimation is the process which improves the compression efficiency of these coding standards by determining the minimum distortion between the current frame and the reference frame. For the past two decades, various Motion Estimation algorithms are implemented in hardware and research is still going on for realizing an optimized hardware solution for this critical module. Efficient implementation of ME in hardware is essential for high-resolution video applications such as HDTV to increase the decoding throughput and to achieve high compression ratio. A review and analysis of various hardware architectures of ME used for H.264 and H.265 coding standards is presented in this paper.  


Author(s):  
Harilaos Koumaras ◽  
Evangellos Pallis ◽  
Anastasios Kourtis ◽  
Drakoulis Martakos

Multimedia applications and services have already possessed a major portion of today’s traffic over communication networks. The revolution and evolution of the World Wide Web has enabled the wide provision of multimedia content over the Internet and any other autonomous network. Among the various types of multimedia, video services (transmission of moving images and sound) are proven dominant for present and future communication networks. Although the available network bandwidth and the corresponding supporting bit rates continue to increase, the raw video data necessitate high bandwidth requirements for its transmission. For example, current commercial communication networks throughput rates are insufficient to handle raw video in real time, even if low spatial and temporal resolution (i.e., frame size and frame rate) has been selected. Towards alleviating the network bandwidth requirements for efficient transmission of audiovisual content, coding techniques have been applied on raw video data, which perform compression by exploiting both temporal and spatial redundancy in video sequences. Video coding is defined as the process of compressing and decompressing a raw digital video sequence, which results in lower data volumes, besides enabling the transmission of video signals over bandwidth-limited means, where uncompressed video signals would not be possible to be transmitted. The use of coding and compression techniques leads to better exploitation and more efficient management of the available bandwidth. Video compression algorithms exploit the fact that a video signal consists of sequence series with high similarity in the spatial, temporal, and frequency domain. Thus, by removing this redundancy in these three different domain types, it is possible to achieve high compression of the deduced data, sacrificing a certain amount of visual information, which however it is not highly noticeable by the mechanisms of the human visual system, which is not sensitive at this type of visual degradation (Richardson, 2003). Thus, the research area of video compression has been a very active field during the last few years by proposing various algorithms and techniques for video coding (International Telecommunications Union [ITU], 1993; ITU 2005a, 2005b; Moving Picture Experts Group [MPEG], 1998; MPEG, 2005a, 2005b). In general video compression techniques can be classified into two classes: (1) the lossy ones and (2) information preserving (lossless). The first methods, although maintaining the video quality of the original/uncompressed signal, do not succeed high compression ratios, while the lossless ones compress more efficiently the data volume of initial raw video signal with the cost of degrading the perceived quality of the video service. The lossy video coding techniques are widely used, in contrast to lossless ones, due to their better performance. More specifically, by enhancing the encoding algorithms and techniques, the latest proposed coding methods try to perform in a more efficient way both the data compression and the maintenance of the deduced perceived quality of the encoded signal at high levels. In this framework, many of these coding techniques and algorithms have been standardized, encouraging by this way the interoperability between various products designed and developed by different manufacturers. This article deals with the fundamentals of the video coding process of the lossy encoding techniques that are common on the great majority of today’s video coding standards and techniques.


Author(s):  
Kok Keong ◽  
Myo Tun ◽  
Yoong Choon Chang

Dirac was started off by British Broadcasting Corp. (BBC) in 2003 as an experimental video coding system based on wavelet technology, which is different from that used in the main proprietary/standard video compression systems. Over the years, Dirac has grown out of its initial development and it is now on offer as an advanced royalty-free video coding system designed for a wide range of users, from delivering low-resolution web content to broadcasting high-definition (HD) and beyond, to near-lossless studio editing. The Dirac’s video coding architecture and algorithms are designed with the “keep it simple” mindset. In spite of that the Dirac seems to give a two-fold reduction in bitrate over MPEG-2 for HD video and broadly competitive with state-of-the-art video codecs. This chapter introduces the architecture of Dirac video encoder. The overall encoding structure is discussed followed by the detail description of motion estimation, Overlapped Block-based Motion Compensation (OBMC), Discrete Wavelet Transform (DWT), Rate Distortion Optimization (RDO) quantization and entropy coding. The Dirac’s bitstream syntax for compressed video data storage and streaming is described. Besides that, the coding performance of Dirac in terms of compression ratio, PSNR, SSIM and VQM in comparison with H.264 as a reference are discussed. Related issues such as transcoding and streaming over packat erasure channel are also discussed.


Dynamic Adaptive Streaming over HTTP (DASH) is an emerging solution that aims to standardize existing proprietary streaming systems. DASH specification defines the media presentation description (MPD), which describes a list of available content, URL addresses, and the segment format. High bandwidth demands in interactive streaming applications pose challenges in efficiently utilizing the available bandwidth. In this paper, a novel Relative Strength Index (RSI) with Geometric mean (GM) namely RSI-GM is proposed for estimating available bandwidth for DASH. The proposed work starts by taking the video as an input at the transmitter side and then the video compression is performed using the TRLE. Then MD5 hashing-based AES encryption is applied to the compressed video data to provide data security. Then RSI-GM is proposed to estimate the available bandwidth for DASH. Finally, after estimation, the bitrate for estimated bandwidth is selected optimally using the Improved Shark Smell Optimization (ISSO) algorithm.


Sign in / Sign up

Export Citation Format

Share Document