Ensemble perception in depth: Correct size-distance rescaling of multiple objects before averaging.

2019 ◽  
Vol 148 (4) ◽  
pp. 728-738 ◽  
Author(s):  
Natalia A. Tiurina ◽  
Igor S. Utochkin
2020 ◽  
Author(s):  
Vladislav Khvostov ◽  
Yuri Markov ◽  
Timothy F. Brady ◽  
Igor Utochkin

Many studies have shown that people can rapidly and efficiently categorize the animacy of individual objects and scenes, even with few visual features available. Does this necessarily mean that the visual system has an unlimited capacity to process animacy across the entire visual field? We tested this in an ensemble task requiring observers to judge the relative numerosity of animate vs. inanimate items in briefly presented sets of multiple objects. We generated a set of morphed “animacy continua” between pairs of animal and inanimate object silhouettes and tested them in both individual object categorization and ensemble enumeration. For the ensemble task, we manipulated the ratio between animate and inanimate items present in the display and we also presented two types of animacy distributions: “segmentable” (including only definitely animate and definitely inanimate items) or “non-segmentable” (middle-value, ambiguous morphs pictures were shown along with the definite “extremes”). Our results showed that observers failed to integrate animacy information from multiple items, as they showed very poor performance in the ensemble task and were not sensitive to the distribution type despite their categorization rate for individual objects being near 100%. A control condition using the same design with color as a category-defining dimension elicited both good individual object and ensemble categorization performance and a strong effect of the segmentability type. We conclude that good individual categorization does not necessarily allow people to build ensemble animacy representations, thus showing the limited capacity of animacy perception.


2020 ◽  
Vol 7 (3) ◽  
pp. 4-24
Author(s):  
Aleksei Iakovlev ◽  
◽  
Natalia Tiurina ◽  
Igor Utochkin

Ensemble perception refers to the ability of an observer to precisely estimate summary statistics of multiple objects (average, range, numerosity, etc.) at a glance. This article reviews the properties and research methodology of ensemble perception. Further, we consider the theoretical debate around mechanisms of information sampling and summary statistics calculation. One theory suggests a coarse, parallel and exhaustive mechanism, whereas another theory assumes high-precision processing of a small subsample of items to accomplish proxy statistics for the entire ensemble. We describe the evolving view of the internal ensemble representation that initially was viewed as a single magnitude (e.g., average) but later thought of as the entire feature distribution of all items. We also discuss the role of ensemble representations in various perceptual tasks. Finally, we describe potential neural correlates and neurally plausible models of ensemble perception.


Author(s):  
J.R. McIntosh ◽  
D.L. Stemple ◽  
William Bishop ◽  
G.W. Hannaway

EM specimens often contain 3-dimensional information that is lost during micrography on a single photographic film. Two images of one specimen at appropriate orientations give a stereo view, but complex structures composed of multiple objects of graded density that superimpose in each projection are often difficult to decipher in stereo. Several analytical methods for 3-D reconstruction from multiple images of a serially tilted specimen are available, but they are all time-consuming and computationally intense.


2019 ◽  
Vol 5 (1) ◽  
pp. 1-9
Author(s):  
Piotr Gulgowski

Abstract Singular nouns in the scope of a distributive operator have been shown to be treated as conceptually plural (Patson and Warren, 2010). The source of this conceptual plurality is not fully clear. In particular, it is not known whether the concept of plurality associated with a singular noun originates from distributing over multiple objects or multiple events. In the present experiment, iterative expressions (distribution over events) were contrasted with collective and distributive sentences using a Stroop-like interference technique (Berent, Pinker, Tzelgov, Bibi, and Goldfarb, 2005; Patson and Warren, 2010). A trend in the data suggests that event distributivity does not elicit a plural interpretation of a grammatically singular noun, however the results were not statistically significant. Possible causes of the non-significant results are discussed.


Author(s):  
Wei Huang ◽  
Xiaoshu Zhou ◽  
Mingchao Dong ◽  
Huaiyu Xu

AbstractRobust and high-performance visual multi-object tracking is a big challenge in computer vision, especially in a drone scenario. In this paper, an online Multi-Object Tracking (MOT) approach in the UAV system is proposed to handle small target detections and class imbalance challenges, which integrates the merits of deep high-resolution representation network and data association method in a unified framework. Specifically, while applying tracking-by-detection architecture to our tracking framework, a Hierarchical Deep High-resolution network (HDHNet) is proposed, which encourages the model to handle different types and scales of targets, and extract more effective and comprehensive features during online learning. After that, the extracted features are fed into different prediction networks for interesting targets recognition. Besides, an adjustable fusion loss function is proposed by combining focal loss and GIoU loss to solve the problems of class imbalance and hard samples. During the tracking process, these detection results are applied to an improved DeepSORT MOT algorithm in each frame, which is available to make full use of the target appearance features to match one by one on a practical basis. The experimental results on the VisDrone2019 MOT benchmark show that the proposed UAV MOT system achieves the highest accuracy and the best robustness compared with state-of-the-art methods.


Author(s):  
Jiahui Huang ◽  
Sheng Yang ◽  
Zishuo Zhao ◽  
Yu-Kun Lai ◽  
Shi-Min Hu

AbstractWe present a practical backend for stereo visual SLAM which can simultaneously discover individual rigid bodies and compute their motions in dynamic environments. While recent factor graph based state optimization algorithms have shown their ability to robustly solve SLAM problems by treating dynamic objects as outliers, their dynamic motions are rarely considered. In this paper, we exploit the consensus of 3D motions for landmarks extracted from the same rigid body for clustering, and to identify static and dynamic objects in a unified manner. Specifically, our algorithm builds a noise-aware motion affinity matrix from landmarks, and uses agglomerative clustering to distinguish rigid bodies. Using decoupled factor graph optimization to revise their shapes and trajectories, we obtain an iterative scheme to update both cluster assignments and motion estimation reciprocally. Evaluations on both synthetic scenes and KITTI demonstrate the capability of our approach, and further experiments considering online efficiency also show the effectiveness of our method for simultaneously tracking ego-motion and multiple objects.


Sign in / Sign up

Export Citation Format

Share Document