scholarly journals Research on Target Object Recognition Based on Transfer-Learning Convolutional SAE in Intelligent Urban Construction

IEEE Access ◽  
2019 ◽  
Vol 7 ◽  
pp. 125357-125368
Author(s):  
Bing Xie ◽  
Zhemin Duan ◽  
Bin Zheng ◽  
Liping Liu
Electronics ◽  
2020 ◽  
Vol 9 (2) ◽  
pp. 210 ◽  
Author(s):  
Yi-Chun Du ◽  
Muslikhin Muslikhin ◽  
Tsung-Han Hsieh ◽  
Ming-Shyan Wang

This paper develops a hybrid algorithm of adaptive network-based fuzzy inference system (ANFIS) and regions with convolutional neural network (R-CNN) for stereo vision-based object recognition and manipulation. The stereo camera at an eye-to-hand configuration firstly captures the image of the target object. Then, the shape, features, and centroid of the object are estimated. Similar pixels are segmented by the image segmentation method, and similar regions are merged through selective search. The eye-to-hand calibration is based on ANFIS to reduce computing burden. A six-degree-of-freedom (6-DOF) robot arm with a gripper will conduct experiments to demonstrate the effectiveness of the proposed system.


Optik ◽  
2014 ◽  
Vol 125 (1) ◽  
pp. 482-485 ◽  
Author(s):  
Zhiping Dan ◽  
Nong Sang ◽  
Yonggang He ◽  
Shuifa Sun

Author(s):  
Michael S. Brickner ◽  
Amir Zvuloni

Thermal imaging (TI) systems, transform the distribution of relative temperatures in a scene into a visible TV image. TIs differ significantly from regular TV images. Most TI systems allow their operators to select preferred polarity which determines the way in which gray shades represent different temperatures. Polarity may be set to either black hot (BH) or white hot (WH). The present experiments were designed to investigate the effects of polarity on object recognition performance in TI and to compare object recognition performance of experts and novices. In the first experiment, twenty flight candidates were asked to recognize target objects in 60 dynamic TI recordings taken from two different TI systems. The targets included a variety of human placed and natural objects. Each subject viewed half the targets in BH and the other half in WH polarity in a balanced experimental design. For 24 out of the 60 targets one direction of polarity produced better performance than the other. Although the direction of superior polarity (BH or WH better) was not consistent, the preferred representation of the target object was very consistent. For example, vegetation was more readily recognized when presented as dark objects on a brighter background. The results are discussed in terms of importance of surface determinants versus edge determinants in the recognition of TI objects. In the second experiment, the performance of 10 expert TI users was found to be significantly more accurate but not much faster than the performance of 20 novice subjects.


2008 ◽  
Vol 20 (3) ◽  
pp. 371-388 ◽  
Author(s):  
Nurit Gronau ◽  
Maital Neta ◽  
Moshe Bar

Visual context plays a prominent role in everyday perception. Contextual information can facilitate recognition of objects within scenes by providing predictions about objects that are most likely to appear in a specific setting, along with the locations that are most likely to contain objects in the scene. Is such identity-related (“semantic”) and location-related (“spatial”) contextual knowledge represented separately or jointly as a bound representation? We conducted a functional magnetic resonance imaging (fMRI) priming experiment whereby semantic and spatial contextual relations between prime and target object pictures were independently manipulated. This method allowed us to determine whether the two contextual factors affect object recognition with or without interacting, supporting a unified versus independent representations, respectively. Results revealed a Semantic × Spatial interaction in reaction times for target object recognition. Namely, significant semantic priming was obtained when targets were positioned in expected (congruent), but not in unexpected (incongruent), locations. fMRI results showed corresponding interactive effects in brain regions associated with semantic processing (inferior prefrontal cortex), visual contextual processing (parahippocampal cortex), and object-related processing (lateral occipital complex). In addition, activation in fronto-parietal areas suggests that attention and memory-related processes might also contribute to the contextual effects observed. These findings indicate that object recognition benefits from associative representations that integrate information about objects' identities and their locations, and directly modulate activation in object-processing cortical regions. Such context frames are useful in maintaining a coherent and meaningful representation of the visual world, and in providing a platform from which predictions can be generated to facilitate perception and action.


Author(s):  
Shih-Wei Liu ◽  
Jen-Yuan (James) Chang

Abstract With the development of Automation industry, a new industrial model has been born, and traditional human resources have gradually been replaced by machines. The World Economic Forum (WEF) pointed out in “The Future of Jobs Report 2018” that the world is experiencing a “workplace revolution”, which means that machine will play a more important role in the future. In response to this situation, in this paper, techniques for object recognition and tracking on a conveyor using eye-in-hand gripper are presented, which are useful in production line for automatic object classification. The eye-in-hand configuration is the most suitable for camera and gripper application because the camera coordinate is the same as the gripper coordinate. The main advantages of eye-in-hand configuration are as follow: (1) occlusion avoidance (2) intuitive teleoperation (3) image from different angles (4) simple calibration. The main difference with eye-on-hand configuration is that it may be out of view sight when the camera is too close to the object. The experimental result is using the eye-in-hand robotic gripper to establish a tracking system to chase the target object. Preliminary results show that the speed of the conveyor can be calculated and the moving distance between the robot and the object is very close after a period of time. It means that the tracking system is successful.


Electronics ◽  
2020 ◽  
Vol 9 (3) ◽  
pp. 380 ◽  
Author(s):  
Agnese Chiatti ◽  
Gianluca Bardaro ◽  
Emanuele Bastianelli ◽  
Ilaria Tiddi ◽  
Prasenjit Mitra ◽  
...  

To assist humans with their daily tasks, mobile robots are expected to navigate complex and dynamic environments, presenting unpredictable combinations of known and unknown objects. Most state-of-the-art object recognition methods are unsuitable for this scenario because they require that: (i) all target object classes are known beforehand, and (ii) a vast number of training examples is provided for each class. This evidence calls for novel methods to handle unknown object classes, for which fewer images are initially available (few-shot recognition). One way of tackling the problem is learning how to match novel objects to their most similar supporting example. Here, we compare different (shallow and deep) approaches to few-shot image matching on a novel data set, consisting of 2D views of common object types drawn from a combination of ShapeNet and Google. First, we assess if the similarity of objects learned from a combination of ShapeNet and Google can scale up to new object classes, i.e., categories unseen at training time. Furthermore, we show how normalising the learned embeddings can impact the generalisation abilities of the tested methods, in the context of two novel configurations: (i) where the weights of a Convolutional two-branch Network are imprinted and (ii) where the embeddings of a Convolutional Siamese Network are L2-normalised.


2019 ◽  
Vol 35 (4) ◽  
pp. 987-998 ◽  
Author(s):  
Pietro Falco ◽  
Shuang Lu ◽  
Ciro Natale ◽  
Salvatore Pirozzi ◽  
Dongheui Lee

Sign in / Sign up

Export Citation Format

Share Document