scholarly journals Color-based estimates of stimulus similarity predict perceptual similarity of image pairs to monkeys

2010 ◽  
Vol 3 (9) ◽  
pp. 511-511 ◽  
Author(s):  
S. R Allred ◽  
J. Y Skiver ◽  
B. Jagadeesh

Author(s):  
Hadar Ram ◽  
Dieter Struyf ◽  
Bram Vervliet ◽  
Gal Menahem ◽  
Nira Liberman

Abstract. People apply what they learn from experience not only to the experienced stimuli, but also to novel stimuli. But what determines how widely people generalize what they have learned? Using a predictive learning paradigm, we examined the hypothesis that a low (vs. high) probability of an outcome following a predicting stimulus would widen generalization. In three experiments, participants learned which stimulus predicted an outcome (S+) and which stimulus did not (S−) and then indicated how much they expected the outcome after each of eight novel stimuli ranging in perceptual similarity to S+ and S−. The stimuli were rings of different sizes and the outcome was a picture of a lightning bolt. As hypothesized, a lower probability of the outcome widened generalization. That is, novel stimuli that were similar to S+ (but not to S−) produced expectations for the outcome that were as high as those associated with S+.









2020 ◽  
Vol 2020 (8) ◽  
pp. 114-1-114-7
Author(s):  
Bryan Blakeslee ◽  
Andreas Savakis

Change detection in image pairs has traditionally been a binary process, reporting either “Change” or “No Change.” In this paper, we present LambdaNet, a novel deep architecture for performing pixel-level directional change detection based on a four class classification scheme. LambdaNet successfully incorporates the notion of “directional change” and identifies differences between two images as “Additive Change” when a new object appears, “Subtractive Change” when an object is removed, “Exchange” when different objects are present in the same location, and “No Change.” To obtain pixel annotated change maps for training, we generated directional change class labels for the Change Detection 2014 dataset. Our tests illustrate that LambdaNet would be suitable for situations where the type of change is unstructured, such as change detection scenarios in satellite imagery.



2020 ◽  
Vol 64 (2) ◽  
pp. 20506-1-20506-7
Author(s):  
Min Zhu ◽  
Rongfu Zhang ◽  
Pei Ma ◽  
Xuedian Zhang ◽  
Qi Guo

Abstract Three-dimensional (3D) reconstruction is extensively used in microscopic applications. Reducing excessive error points and achieving accurate matching of weak texture regions have been the classical challenges for 3D microscopic vision. A Multi-ST algorithm was proposed to improve matching accuracy. The process is performed in two main stages: scaled microscopic images and regularized cost aggregation. First, microscopic image pairs with different scales were extracted according to the Gaussian pyramid criterion. Second, a novel cost aggregation approach based on the regularized multi-scale model was implemented into all scales to obtain the final cost. To evaluate the performances of the proposed Multi-ST algorithm and compare different algorithms, seven groups of images from the Middlebury dataset and four groups of experimental images obtained by a binocular microscopic system were analyzed. Disparity maps and reconstruction maps generated by the proposed approach contained more information and fewer outliers or artifacts. Furthermore, 3D reconstruction of the plug gauges using the Multi-ST algorithm showed that the error was less than 0.025 mm.



2020 ◽  
Vol 12 (3) ◽  
pp. 371 ◽  
Author(s):  
Sahar Dehnavi ◽  
Yasser Maghsoudi ◽  
Klemen Zakšek ◽  
Mohammad Javad Valadan Zoej ◽  
Gunther Seckmeyer ◽  
...  

Due to the considerable impact of clouds on the energy balance in the atmosphere and on the earth surface, they are of great importance for various applications in meteorology or remote sensing. An important aspect of the cloud research studies is the detection of cloudy pixels from the processing of satellite images. In this research, we investigated a stereographic method on a new set of Meteosat images, namely the combination of the high resolution visible (HRV) channel of the Meteosat-8 Indian Ocean Data Coverage (IODC) as a stereo pair with the HRV channel of the Meteosat Second Generation (MSG) Meteosat-10 image at 0° E. In addition, an approach based on the outputs from stereo analysis was proposed to detect cloudy pixels. This approach is introduced with a 2D-scatterplot based on the parallax value and the minimum intersection distance. The mentioned scatterplot was applied to determine/detect cloudy pixels in various image subsets with different amounts of cloud cover. Apart from the general advantage of the applied stereography method, which only depends on geometric relationships, the cloud detection results are also improved because: (1) The stereo pair is the HRV bands of the Spinning Enhanced Visible and InfraRed Imager (SEVIRI) sensor, with the highest spatial resolution available from the Meteosat geostationary platform; and (2) the time difference between the image pairs is nearly 5 s, which improves the matching results and also decreases the effect of cloud movements. In order to prove this improvement, the results of this stereo-based approach were compared with three different reflectance-based target detection techniques, including the adaptive coherent estimator (ACE), constrained energy minimization (CEM), and matched filter (MF). The comparison of the receiver operating characteristics (ROC) detection curves and the area under these curves (AUC) showed better detection results with the proposed method. The AUC value was 0.79, 0.90, 0.90, and 0.93 respectively for ACE, CEM, MF, and the proposed stereo-based detection approach. The results of this research shall enable a more realistic modelling of down-welling solar irradiance in the future.



Author(s):  
Jin Zhou ◽  
Qing Zhang ◽  
Jian-Hao Fan ◽  
Wei Sun ◽  
Wei-Shi Zheng

AbstractRecent image aesthetic assessment methods have achieved remarkable progress due to the emergence of deep convolutional neural networks (CNNs). However, these methods focus primarily on predicting generally perceived preference of an image, making them usually have limited practicability, since each user may have completely different preferences for the same image. To address this problem, this paper presents a novel approach for predicting personalized image aesthetics that fit an individual user’s personal taste. We achieve this in a coarse to fine manner, by joint regression and learning from pairwise rankings. Specifically, we first collect a small subset of personal images from a user and invite him/her to rank the preference of some randomly sampled image pairs. We then search for the K-nearest neighbors of the personal images within a large-scale dataset labeled with average human aesthetic scores, and use these images as well as the associated scores to train a generic aesthetic assessment model by CNN-based regression. Next, we fine-tune the generic model to accommodate the personal preference by training over the rankings with a pairwise hinge loss. Experiments demonstrate that our method can effectively learn personalized image aesthetic preferences, clearly outperforming state-of-the-art methods. Moreover, we show that the learned personalized image aesthetic benefits a wide variety of applications.



Author(s):  
Bin Wang ◽  
Dapeng Tao ◽  
Rui Dong ◽  
Yuanyan Tang ◽  
Xinbo Gao
Keyword(s):  


Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1299
Author(s):  
Honglin Yuan ◽  
Tim Hoogenkamp ◽  
Remco C. Veltkamp

Deep learning has achieved great success on robotic vision tasks. However, when compared with other vision-based tasks, it is difficult to collect a representative and sufficiently large training set for six-dimensional (6D) object pose estimation, due to the inherent difficulty of data collection. In this paper, we propose the RobotP dataset consisting of commonly used objects for benchmarking in 6D object pose estimation. To create the dataset, we apply a 3D reconstruction pipeline to produce high-quality depth images, ground truth poses, and 3D models for well-selected objects. Subsequently, based on the generated data, we produce object segmentation masks and two-dimensional (2D) bounding boxes automatically. To further enrich the data, we synthesize a large number of photo-realistic color-and-depth image pairs with ground truth 6D poses. Our dataset is freely distributed to research groups by the Shape Retrieval Challenge benchmark on 6D pose estimation. Based on our benchmark, different learning-based approaches are trained and tested by the unified dataset. The evaluation results indicate that there is considerable room for improvement in 6D object pose estimation, particularly for objects with dark colors, and photo-realistic images are helpful in increasing the performance of pose estimation algorithms.



Sign in / Sign up

Export Citation Format

Share Document