scholarly journals Inverse Tone Mapping Operator Using Sequential Deep Neural Networks Based on the Human Visual System

IEEE Access ◽  
2018 ◽  
Vol 6 ◽  
pp. 52058-52072 ◽  
Author(s):  
Hanbyol Jang ◽  
Kihun Bang ◽  
Jinseong Jang ◽  
Dosik Hwang
IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 118359-118369
Author(s):  
Disheng Miao ◽  
Zhongjie Zhu ◽  
Yongqiang Bai ◽  
Gangyi Jiang ◽  
Zhiyong Duan

Algorithms ◽  
2020 ◽  
Vol 13 (7) ◽  
pp. 167 ◽  
Author(s):  
Dan Malowany ◽  
Hugo Guterman

Computer vision is currently one of the most exciting and rapidly evolving fields of science, which affects numerous industries. Research and development breakthroughs, mainly in the field of convolutional neural networks (CNNs), opened the way to unprecedented sensitivity and precision in object detection and recognition tasks. Nevertheless, the findings in recent years on the sensitivity of neural networks to additive noise, light conditions, and to the wholeness of the training dataset, indicate that this technology still lacks the robustness needed for the autonomous robotic industry. In an attempt to bring computer vision algorithms closer to the capabilities of a human operator, the mechanisms of the human visual system was analyzed in this work. Recent studies show that the mechanisms behind the recognition process in the human brain include continuous generation of predictions based on prior knowledge of the world. These predictions enable rapid generation of contextual hypotheses that bias the outcome of the recognition process. This mechanism is especially advantageous in situations of uncertainty, when visual input is ambiguous. In addition, the human visual system continuously updates its knowledge about the world based on the gaps between its prediction and the visual feedback. CNNs are feed forward in nature and lack such top-down contextual attenuation mechanisms. As a result, although they process massive amounts of visual information during their operation, the information is not transformed into knowledge that can be used to generate contextual predictions and improve their performance. In this work, an architecture was designed that aims to integrate the concepts behind the top-down prediction and learning processes of the human visual system with the state-of-the-art bottom-up object recognition models, e.g., deep CNNs. The work focuses on two mechanisms of the human visual system: anticipation-driven perception and reinforcement-driven learning. Imitating these top-down mechanisms, together with the state-of-the-art bottom-up feed-forward algorithms, resulted in an accurate, robust, and continuously improving target recognition model.


2021 ◽  
Author(s):  
Nipu Rani Barai

With the growing popularity of High Dynamic Range Imaging (HDRI), the necessity for advanced tone-mapping techniques has greatly increased. In this thesis, I propose a novel saliency guided edge-preserving tone-mapping method that uses saliency region information of an HDR image as input to a guided filter for base and detail image layer separation. Both high resolution and low resolution saliency maps were used for the performance evaluation of the proposed method. After detail layer enhancement and base layer compression with constant weights, a new edge preserved tone-mapped image was composed by adding the layers back together with saturation and exposure adjustments. The filter operation is faster due to the use of the guided filter, which has O(N) time operation with N number of pixels. Both objective and subjective quality assessment results demonstrated that the proposed method has higher edge and naturalness preserving capability, which is homologous to the Human Visual System (HVS), as compared to other state-of-the-art tone-mapping approaches.


2006 ◽  
pp. 187-221 ◽  
Author(s):  
Erik Reinhard ◽  
Greg Ward ◽  
Sumanta Pattanaik ◽  
Paul Debevec

2004 ◽  
Author(s):  
Alessandro Rizzi ◽  
Carlo Gatta ◽  
Benedetta Piacentini ◽  
Massimo Fierro ◽  
Daniele Marini

2017 ◽  
Author(s):  
H. Steven Scholte ◽  
Max M. Losch ◽  
Kandan Ramakrishnan ◽  
Edward H.F. de Haan ◽  
Sander M. Bohte

AbstractVision research has been shaped by the seminal insight that we can understand the higher-tier visual cortex from the perspective of multiple functional pathways with different goals. In this paper, we try to give a computational account of the functional organization of this system by reasoning from the perspective of multi-task deep neural networks. Machine learning has shown that tasks become easier to solve when they are decomposed into subtasks with their own cost function. We hypothesize that the visual system optimizes multiple cost functions of unrelated tasks and this causes the emergence of a ventral pathway dedicated to vision for perception, and a dorsal pathway dedicated to vision for action. To evaluate the functional organization in multi-task deep neural networks, we propose a method that measures the contribution of a unit towards each task, applying it to two networks that have been trained on either two related or two unrelated tasks, using an identical stimulus set. Results show that the network trained on the unrelated tasks shows a decreasing degree of feature representation sharing towards higher-tier layers while the network trained on related tasks uniformly shows high degree of sharing. We conjecture that the method we propose can be used to analyze the anatomical and functional organization of the visual system and beyond. We predict that the degree to which tasks are related is a good descriptor of the degree to which they share downstream cortical-units.


Sign in / Sign up

Export Citation Format

Share Document