A Multi-Scale Cascaded Hierarchical Model for Image Labeling

Author(s):  
Degui Xiao ◽  
Qilei Chen ◽  
Shanshan Li

Image labeling is an important and challenging task in the area of graphics and visual computing, where datasets with high quality labeling are critically needed. In this paper, based on the commonly accepted observation that the same semantic object in images with different resolutions may have different representations, we propose a novel multi-scale cascaded hierarchical model (MCHM) to enhance general image labeling methods. Our proposed approach first creates multi-resolution images from the original one to form an image pyramid and labels each image at different scale individually. Next, it constructs a cascaded hierarchical model and a feedback circle between image pyramid and labeling methods. The original image labeling result is used to adjust labeling parameters of those scaled images. Labeling results from the scaled images are then fed back to enhance the original image labeling results. These naturally form a global optimization problem under scale-space condition. We further propose a desirable iterative algorithm in order to run the model. The global convergence of the algorithm is proven through iterative approximation with latent optimization constraints. We have conducted extensive experiments with five widely used labeling methods on five popular image datasets. Experimental results indicate that MCHM improves labeling accuracy of the state-of-the-art image labeling approaches impressively.

2007 ◽  
Vol 28 (5) ◽  
pp. 545-554 ◽  
Author(s):  
Xiaohong Zhang ◽  
Ming Lei ◽  
Dan Yang ◽  
Yuzhu Wang ◽  
Litao Ma

2016 ◽  
Vol 32 (8) ◽  
pp. 1709-1720 ◽  
Author(s):  
J. A. Salo ◽  
D. M. Theobald

2007 ◽  
Vol 33 (4) ◽  
pp. 414-417 ◽  
Author(s):  
Yu-Zhu WANG ◽  
Dan YANG ◽  
Xiao-Hong ZHANG

2003 ◽  
Vol 7 (1_suppl) ◽  
pp. 125-155
Author(s):  
Maja Serman ◽  
Niall J. L. Griffith

In this paper we approach the subject of modelling and understanding segmentation processes in melodic perception using a temporal multi-scale representation framework. We start with the hypothesis that segmentation depends on the ability of the perceptual system to detect changes in the sensory signal. In particular, we are interested in a model of change detection in music perception that would help us to investigate functional aspects of low-level perceptual processes in music and their universality in terms of the general properties of the auditory system. To investigate this hypothesis, we have developed a temporal multi-scale model that mimics the ability of the listener to detect changes in pitch, loudness and timbre when listening to performed melodies. The model is set within the linear scale-space theoretical framework, as developed for image structure analysis but in this case applied to the temporal processing domain. It is structured in such a way as to enable us to verify the assumption that segmentation is influenced by both the dynamics of signal propagation through a neural map and learning and attention factors. Consequently, the model is examined from two perspectives: 1) the computational architecture which models signal propagation is examined for achieving the effects of the universal, inborn aspects of segmentation 2) the model structure capable of influencing choices of segmentation outcomes is explained and some of its effects are examined in view of the known segmentation results. The results of the presented case studies demonstrate that the model accounts for some effects of perceptual organization of the sensory signal and provides a sound basis for analysing different types of changes and coordination across the melodic descriptors in segmentation decisions.


Perception ◽  
1997 ◽  
Vol 26 (1_suppl) ◽  
pp. 136-136
Author(s):  
J H Elder

There is both psychophysical and physiological evidence that the perception of brightness variations in an image may be the result of a filling-in process in which the luminance signal is encoded only at image contours and is then neurally diffused to form representations of surface brightness. Despite this evidence, the filling-in hypothesis remains controversial. One problem is that in previous experiments highly simplified synthetic stimuli have been used; it is unclear whether brightness filling-in is feasible for complex natural images containing shading, shadows, and focal blur. To address this question, we present a computational model for brightness filling-in and results of experiments which test the model on a large and diverse set of natural images. The model is based on a scale-space method for edge detection which computes a contour code consisting of estimates of position, brightness, contrast, and blur at each edge point in an image (Elder and Zucker, 1996, paper presented at ECCV). This representation is then inverted by a diffusion-based filling-in algorithm which reconstructs an estimate of the original image. Psychophysical assessment of results shows that while filling-in of brightness alone leads to significant artifact, parallel filling-in of both brightness and blur produces perceptually accurate reconstructions. The temporal dynamics of blur reconstruction predicted by the model are consistent with psychophysical studies of blur perception (Westheimer, 1991 Journal of the Optical Society of America A8 681 – 685). These results suggest that a scale-adaptive contour representation can in principle capture the information needed for the perceptually accurate filling-in of complex natural images.


1994 ◽  
Vol 04 (04) ◽  
pp. 467-475 ◽  
Author(s):  
PIERRE-LOUIS LIONS

We briefly review the derivation due to Alvarez, Guichard, Morel and the author of mathematical models in Image Processing. We deduce from classical axions in Computer Vision some nonlinear partial differential equations of evolution type that correspond to general multi-scale analysis (scale-space). We also obtain specific nonlinear models that satisfy additional invariances which are relevant for the analysis of images.


Sign in / Sign up

Export Citation Format

Share Document