scholarly journals Perceptual Video Coding Scheme Using Just Noticeable Distortion Model Based on Entropy Filter

Entropy ◽  
2019 ◽  
Vol 21 (11) ◽  
pp. 1095 ◽  
Author(s):  
Cui ◽  
Peng ◽  
Jiang ◽  
Chen ◽  
Yu

Because perceptual video coding (PVC) can reduce bitrates with negligible visual quality loss in video compression, a PVC scheme based on just noticeable distortion (JND) model is proposed for ultra-high definition video. Firstly, the proposed JND model is designed, considering the spatial JND characteristics such as contrast sensitivity, luminance adaptation and saliency weight factor. Secondly, in order to perform precise JND suppression, the Gauss differential entropy (GDE) filter is designed to divide the image into smooth and complex texture region. Thirdly, through incorporating the proposed JND model into the encoding process, the transform coefficients are suppressed in harmonization with the transform/quantization process of high efficiency video coding (HEVC). In order to achieve the JND suppression effectively, a distortion compensation factor and distortion compensation control factor are incorporated to control the extent of distortion in the rate distortion optimization process. The experimental results show that the proposed PVC scheme can achieve a remarkable bitrate reduction of 32.98% for low delay (LD) configuration and 28.61% for random access (RA) configuration with a negligible subjective quality loss. Meanwhile, the proposed method only causes about average 12.94% and 22.45% encoding time increase under LD and RA configuration compared with an HEVC reference software, respectively.

Author(s):  
Diego Jesus Serrano-Carrasco ◽  
Antonio Jesus Diaz-Honrubia ◽  
Pedro Cuenca

AbstractWith the advent of smartphones and tablets, video traffic on the Internet has increased enormously. With this in mind, in 2013 the High Efficiency Video Coding (HEVC) standard was released with the aim of reducing the bit rate (at the same quality) by 50% with respect to its predecessor. However, new contents with greater resolutions and requirements appear every day, making it necessary to further reduce the bit rate. Perceptual video coding has recently been recognized as a promising approach to achieving high-performance video compression and eye tracking data can be used to create and verify these models. In this paper, we present a new algorithm for the bit rate reduction of screen recorded sequences based on the visual perception of videos. An eye tracking system is used during the recording to locate the fixation point of the viewer. Then, the area around that point is encoded with the base quantization parameter (QP) value, which increases when moving away from it. The results show that up to 31.3% of the bit rate may be saved when compared with the original HEVC-encoded sequence, without a significant impact on the perceived quality.


Author(s):  
Mohammad Barr

Background: High-Efficiency Video Coding (HEVC) is a recent video compression standard. It provides better compression performance compared to its predecessor, H.264/AVC. However, the computational complexity of the HEVC encoder is much higher than that of H.264/AVC encoder. This makes HEVC less attractive to be used in real-time applications and in devices with limited resources (e.g., low memory, low processing power, etc.). The increased computational complexity of HEVC is partly due to its use of a variable size Transform Unit (TU) selection algorithm which successively performs transform operations using transform units of different sizes before selecting the optimal transform unit size. In this paper, a fast transform unit size selection method is proposed to reduce the computational complexity of an HEVC encoder. Methods: Bayesian decision theory is used to predict the size of the TU during encoding. This is done by exploiting the TU size decisions at a previous temporal level and by modeling the relationship between the TU size and the Rate-Distortion (RD) cost values. Results: Simulation results show that the proposed method achieves a reduction of the encoding time of the latest HEVC encoder by 16.21% on average without incurring any noticeable compromise on its compression efficiency. The algorithm also reduces the number of transform operations by 44.98% on average. Conclusion: In this paper, a novel fast TU size selection scheme for HEVC is proposed. The proposed technique outperforms both the latest HEVC reference software, HM 16.0, as well as other state-of-the-art techniques in terms of time-complexity. The compression performance of the proposed technique is comparable to that of HM 16.0.


2019 ◽  
Vol 2019 ◽  
pp. 1-13
Author(s):  
Qiuwen Zhang ◽  
Shuaichao Wei ◽  
Rijian Su

Three-dimensional extension of the high efficiency video coding (3D-HEVC) is an emerging international video compression standard for multiview video system applications. Similar to HEVC, a computationally expensive mode decision is performed using all depth levels and prediction modes to select the least rate-distortion (RD) cost for each coding unit (CU). In addition, new tools and intercomponent prediction techniques have been introduced to 3D-HEVC for improving the compression efficiency of the multiview texture videos. These techniques, despite achieving the highest texture video coding efficiency, involve extremely high-complex procedures, thus limiting 3D-HEVC encoders in practical applications. In this paper, a fast texture video coding method based on motion homogeneity is proposed to reduce 3D-HEVC computational complexity. Because the multiview texture videos instantly represent the same scene at the same time (considering that the optimal CU depth level and prediction modes are highly multiview content dependent), it is not efficient to use all depth levels and prediction modes in 3D-HEVC. The motion homogeneity model of a CU is first studied according to the motion vectors and prediction modes from the corresponding CUs. Based on this model, we present three efficient texture video coding approaches, such as the fast depth level range determination, early SKIP/Merge mode decision, and adaptive motion search range adjustment. Experimental results demonstrate that the proposed overall method can save 56.6% encoding time with only trivial coding efficiency degradation.


2021 ◽  
Vol 7 (2) ◽  
pp. 39
Author(s):  
Miguel O. Martínez-Rach ◽  
Héctor Migallón ◽  
Otoniel López-Granado ◽  
Vicente Galiano ◽  
Manuel P. Malumbres

The audiovisual entertainment industry has entered a race to find the video encoder offering the best Rate/Distortion (R/D) performance for high-quality high-definition video content. The challenge consists in providing a moderate to low computational/hardware complexity encoder able to run Ultra High-Definition (UHD) video formats of different flavours (360°, AR/VR, etc.) with state-of-the-art R/D performance results. It is necessary to evaluate not only R/D performance, a highly important feature, but also the complexity of future video encoders. New coding tools offering a small increase in R/D performance at the cost of greater complexity are being advanced with caution. We performed a detailed analysis of two evolutions of High Efficiency Video Coding (HEVC) video standards, Joint Exploration Model (JEM) and Versatile Video Coding (VVC), in terms of both R/D performance and complexity. The results show how VVC, which represents the new direction of future standards, has, for the time being, sacrificed R/D performance in order to significantly reduce overall coding/decoding complexity.


2021 ◽  
Vol 2021 ◽  
pp. 1-10
Author(s):  
Jinchao Zhao ◽  
Yihan Wang ◽  
Qiuwen Zhang

With the development of broadband networks and high-definition displays, people have higher expectations for the quality of video images, which also brings new requirements and challenges to video coding technology. Compared with H.265/High Efficiency Video Coding (HEVC), the latest video coding standard, Versatile Video Coding (VVC), can save 50%-bit rate while maintaining the same subjective quality, but it leads to extremely high encoding complexity. To decrease the complexity, a fast coding unit (CU) size decision method based on Just Noticeable Distortion (JND) and deep learning is proposed in this paper. Specifically, the hybrid JND threshold model is first designed to distinguish smooth, normal, or complex region. Then, if CU belongs to complex area, the Ultra-Spherical SVM (US-SVM) classifiers are trained for forecasting the best splitting mode. Experimental results illustrate that the proposed method can save about 52.35% coding runtime, which can realize a trade-off between the reduction of computational burden and coding efficiency compared with the latest methods.


Author(s):  
Srinivas Bachu ◽  
N. Ramya Teja

Due to the advancement of multimedia and its requirement of communication over the network, video compression has received much attention among the researchers. One of the popular video codings is scalable video coding, referred to as H.264/AVC standard. The major drawback in the H.264 is that it performs the exhaustive search over the interlayer prediction to gain the best rate-distortion performance. To reduce the computation overhead due to exhaustive search on mode prediction process, this paper presents a new technique for inter prediction mode selection based on the fuzzy holoentropy. This proposed scheme utilizes the pixel values and probabilistic distribution of pixel symbols to decide the mode. The adaptive mode selection is introduced here by analyzing the pixel values of the current block to be coded with those of a motion compensated reference block using fuzzy holoentropy. The adaptively selected mode decision can reduce the computation time without affecting the visual quality of frames. Experimentation of the proposed scheme is evaluated by utilizing five videos, and from the analysis, it is evident that proposed scheme has overall high performance with values of 41.367 dB and 0.992 for PSNR and SSIM respectively.


Author(s):  
MyungJun Kim ◽  
Yung-Lyul Lee

High Efficiency Video Coding (HEVC) uses an 8-point filter and a 7-point filter, which are based on the discrete cosine transform (DCT), for the 1/2-pixel and 1/4-pixel interpolations, respectively. In this paper, discrete sine transform (DST)-based interpolation filters (IF) are proposed. The first proposed DST-based IFs (DST-IFs) use 8-point and 7-point filters for the 1/2-pixel and 1/4-pixel interpolations, respectively. The final proposed DST-IFs use 12-point and 11-point filters for the 1/2-pixel and 1/4-pixel interpolations, respectively. These DST-IF methods are proposed to improve the motion-compensated prediction in HEVC. The 8-point and 7-point DST-IF methods showed average BD-rate reductions of 0.7% and 0.3% in the random access (RA) and low delay B (LDB) configurations, respectively. The 12-point and 11-point DST-IF methods showed average BD-rate reductions of 1.4% and 1.2% in the RA and LDB configurations for the Luma component, respectively.


2020 ◽  
Vol 2020 ◽  
pp. 1-11
Author(s):  
Jinchao Zhao ◽  
Yihan Wang ◽  
Qiuwen Zhang

With the development of technology, the hardware requirement and expectations of user for visual enjoyment are getting higher and higher. The multitype tree (MTT) architecture is proposed by the Joint Video Experts Team (JVET). Therefore, it is necessary to determine not only coding unit (CU) depth but also its split mode in the H.266/Versatile Video Coding (H.266/VVC). Although H.266/VVC achieves significant coding performance on the basis of H.265/High Efficiency Video Coding (H.265/HEVC), it causes significantly coding complexity and increases coding time, where the most time-consuming part is traversal calculation rate-distortion (RD) of CU. To solve these problems, this paper proposes an adaptive CU split decision method based on deep learning and multifeature fusion. Firstly, we develop a texture classification model based on threshold to recognize complex and homogeneous CU. Secondly, if the complex CUs belong to edge CU, a Convolutional Neural Network (CNN) structure based on multifeature fusion is utilized to classify CU. Otherwise, an adaptive CNN structure is used to classify CUs. Finally, the division of CU is determined by the trained network and the parameters of CU. When the complex CUs are split, the above two CNN schemes can successfully process the training samples and terminate the rate-distortion optimization (RDO) calculation for some CUs. The experimental results indicate that the proposed method reduces the computational complexity and saves 39.39% encoding time, thereby achieving fast encoding in H.266/VVC.


2020 ◽  
Vol 34 (07) ◽  
pp. 11580-11587
Author(s):  
Haojie Liu ◽  
Han Shen ◽  
Lichao Huang ◽  
Ming Lu ◽  
Tong Chen ◽  
...  

Traditional video compression technologies have been developed over decades in pursuit of higher coding efficiency. Efficient temporal information representation plays a key role in video coding. Thus, in this paper, we propose to exploit the temporal correlation using both first-order optical flow and second-order flow prediction. We suggest an one-stage learning approach to encapsulate flow as quantized features from consecutive frames which is then entropy coded with adaptive contexts conditioned on joint spatial-temporal priors to exploit second-order correlations. Joint priors are embedded in autoregressive spatial neighbors, co-located hyper elements and temporal neighbors using ConvLSTM recurrently. We evaluate our approach for the low-delay scenario with High-Efficiency Video Coding (H.265/HEVC), H.264/AVC and another learned video compression method, following the common test settings. Our work offers the state-of-the-art performance, with consistent gains across all popular test sequences.


Sign in / Sign up

Export Citation Format

Share Document