Learning based 3D keypoint detection with local and global attributes in multi-scale space

Author(s):  
Xinyu Lin ◽  
Ce Zhu ◽  
Qian Zhang ◽  
Mengxue Wang ◽  
Yipeng Liu
Sensors ◽  
2021 ◽  
Vol 21 (7) ◽  
pp. 2280
Author(s):  
Ching-Chang Wong ◽  
Li-Yu Yeh ◽  
Chih-Cheng Liu ◽  
Chi-Yi Tsai ◽  
Hisasuki Aoyama

In this paper, a manipulation planning method for object re-orientation based on semantic segmentation keypoint detection is proposed for robot manipulator which is able to detect and re-orientate the randomly placed objects to a specified position and pose. There are two main parts: (1) 3D keypoint detection system; and (2) manipulation planning system for object re-orientation. In the 3D keypoint detection system, an RGB-D camera is used to obtain the information of the environment and can generate 3D keypoints of the target object as inputs to represent its corresponding position and pose. This process simplifies the 3D model representation so that the manipulation planning for object re-orientation can be executed in a category-level manner by adding various training data of the object in the training phase. In addition, 3D suction points in both the object’s current and expected poses are also generated as the inputs of the next operation stage. During the next stage, Mask Region-Convolutional Neural Network (Mask R-CNN) algorithm is used for preliminary object detection and object image. The highest confidence index image is selected as the input of the semantic segmentation system in order to classify each pixel in the picture for the corresponding pack unit of the object. In addition, after using a convolutional neural network for semantic segmentation, the Conditional Random Fields (CRFs) method is used to perform several iterations to obtain a more accurate result of object recognition. When the target object is segmented into the pack units of image process, the center position of each pack unit can be obtained. Then, a normal vector of each pack unit’s center points is generated by the depth image information and pose of the object, which can be obtained by connecting the center points of each pack unit. In the manipulation planning system for object re-orientation, the pose of the object and the normal vector of each pack unit are first converted into the working coordinate system of the robot manipulator. Then, according to the current and expected pose of the object, the spherical linear interpolation (Slerp) algorithm is used to generate a series of movements in the workspace for object re-orientation on the robot manipulator. In addition, the pose of the object is adjusted on the z-axis of the object’s geodetic coordinate system based on the image features on the surface of the object, so that the pose of the placed object can approach the desired pose. Finally, a robot manipulator and a vacuum suction cup made by the laboratory are used to verify that the proposed system can indeed complete the planned task of object re-orientation.


2007 ◽  
Vol 28 (5) ◽  
pp. 545-554 ◽  
Author(s):  
Xiaohong Zhang ◽  
Ming Lei ◽  
Dan Yang ◽  
Yuzhu Wang ◽  
Litao Ma

2007 ◽  
Vol 33 (4) ◽  
pp. 414-417 ◽  
Author(s):  
Yu-Zhu WANG ◽  
Dan YANG ◽  
Xiao-Hong ZHANG

2003 ◽  
Vol 7 (1_suppl) ◽  
pp. 125-155
Author(s):  
Maja Serman ◽  
Niall J. L. Griffith

In this paper we approach the subject of modelling and understanding segmentation processes in melodic perception using a temporal multi-scale representation framework. We start with the hypothesis that segmentation depends on the ability of the perceptual system to detect changes in the sensory signal. In particular, we are interested in a model of change detection in music perception that would help us to investigate functional aspects of low-level perceptual processes in music and their universality in terms of the general properties of the auditory system. To investigate this hypothesis, we have developed a temporal multi-scale model that mimics the ability of the listener to detect changes in pitch, loudness and timbre when listening to performed melodies. The model is set within the linear scale-space theoretical framework, as developed for image structure analysis but in this case applied to the temporal processing domain. It is structured in such a way as to enable us to verify the assumption that segmentation is influenced by both the dynamics of signal propagation through a neural map and learning and attention factors. Consequently, the model is examined from two perspectives: 1) the computational architecture which models signal propagation is examined for achieving the effects of the universal, inborn aspects of segmentation 2) the model structure capable of influencing choices of segmentation outcomes is explained and some of its effects are examined in view of the known segmentation results. The results of the presented case studies demonstrate that the model accounts for some effects of perceptual organization of the sensory signal and provides a sound basis for analysing different types of changes and coordination across the melodic descriptors in segmentation decisions.


2016 ◽  
Vol 18 (9) ◽  
pp. 1762-1771 ◽  
Author(s):  
Maxim Karpushin ◽  
Giuseppe Valenzise ◽  
Frederic Dufaux

Sign in / Sign up

Export Citation Format

Share Document