scholarly journals Pedestrian Reidentification Based on Feature Fusion and Metric Learning

Author(s):  
Hongbin Luo

The pedestrian recognition in public environment is influenced by the pedestrian environment and the dynamic characteristic boundary factors, so it is easy to produce the tracking error. In order to improve the ability of pedestrian re-identification in public environment, we need to carry out feature fusion and metric learning, and propose pedestrian re-identification based on feature fusion and metric learning. The geometric grid area model of pedestrian recognition in public environment is constructed, the method of fuzzy dynamic feature segmentation is used to reconstruct the dynamic boundary feature point of pedestrian recognition in public environment, the method of bottom-up modeling is used to design the dynamic area grid model of pedestrian recognition in public environment, the design of dynamic area grid model is three-dimensional grid area, the grayscale pixel set of pedestrian recognition dynamic constraint under public environment is extracted, the boundary feature fusion is carried out according to the distribution intensity of grayscale, the image fusion and enhancement information processing of pedestrian recognition under public environment, and the method of 3D dynamic constraint is used to realize the local motion planning of pedestrian recognition under public environment, and the recognition feature fusion and learning of pedestrian recognition under public environment is realized according to the result of contour segmentation. The simulation results show that the method is used for pedestrian recognition again in public environment, and the fuzzy judgment ability of pedestrian dynamic edge features is strong, which makes the error controlled below 10 mm, and the fluctuation of pedestrian recognition again is more stable, the recognition accuracy is higher and the robustness is better.

2019 ◽  
Vol 63 (5) ◽  
pp. 50402-1-50402-9 ◽  
Author(s):  
Ing-Jr Ding ◽  
Chong-Min Ruan

Abstract The acoustic-based automatic speech recognition (ASR) technique has been a matured technique and widely seen to be used in numerous applications. However, acoustic-based ASR will not maintain a standard performance for the disabled group with an abnormal face, that is atypical eye or mouth geometrical characteristics. For governing this problem, this article develops a three-dimensional (3D) sensor lip image based pronunciation recognition system where the 3D sensor is efficiently used to acquire the action variations of the lip shapes of the pronunciation action from a speaker. In this work, two different types of 3D lip features for pronunciation recognition are presented, 3D-(x, y, z) coordinate lip feature and 3D geometry lip feature parameters. For the 3D-(x, y, z) coordinate lip feature design, 18 location points, each of which has 3D-sized coordinates, around the outer and inner lips are properly defined. In the design of 3D geometry lip features, eight types of features considering the geometrical space characteristics of the inner lip are developed. In addition, feature fusion to combine both 3D-(x, y, z) coordinate and 3D geometry lip features is further considered. The presented 3D sensor lip image based feature evaluated the performance and effectiveness using the principal component analysis based classification calculation approach. Experimental results on pronunciation recognition of two different datasets, Mandarin syllables and Mandarin phrases, demonstrate the competitive performance of the presented 3D sensor lip image based pronunciation recognition system.


Sensors ◽  
2019 ◽  
Vol 19 (22) ◽  
pp. 4987 ◽  
Author(s):  
Xu ◽  
Zhang ◽  
Cao ◽  
Pang ◽  
Sun

The three-dimensional (3D) path following problem of an underactuated autonomous underwater vehicle with ocean currents disturbances is addressed in this paper. Firstly, the motion equation under the ocean currents disturbance is established, and the dynamic model of 3D tracking error is constructed based on virtual guidance method. Then, a finite-time control scheme based on super-twisting observer and command filtered backstepping technology is proposed. We adopt super-twisting observer based on finite-time theory to observe the ocean currents disturbances for improving the system robust. A command filtered backstepping is proposed to replace the differential process in the conventional backstepping method for avoiding the differential expansion problem. The filter compensation loop is designed to ensure the accuracy of the filtered signal, and the anti-integration saturation link is designed considering the influence of integral saturation. Lyapunov stability theory is used to prove the stability of the underactuated AUV. Simulation studies are conducted to show the effectiveness and robustness of the controller.


Sensors ◽  
2020 ◽  
Vol 20 (6) ◽  
pp. 1652 ◽  
Author(s):  
Peida Wu ◽  
Ziguan Cui ◽  
Zongliang Gan ◽  
Feng Liu

In recent years, deep learning methods have been widely used in the hyperspectral image (HSI) classification tasks. Among them, spectral-spatial combined methods based on the three-dimensional (3-D) convolution have shown good performance. However, because of the three-dimensional convolution, increasing network depth will result in a dramatic rise in the number of parameters. In addition, the previous methods do not make full use of spectral information. They mostly use the data after dimensionality reduction directly as the input of networks, which result in poor classification ability in some categories with small numbers of samples. To address the above two issues, in this paper, we designed an end-to-end 3D-ResNeXt network which adopts feature fusion and label smoothing strategy further. On the one hand, the residual connections and split-transform-merge strategy can alleviate the declining-accuracy phenomenon and decrease the number of parameters. We can adjust the hyperparameter cardinality instead of the network depth to extract more discriminative features of HSIs and improve the classification accuracy. On the other hand, in order to improve the classification accuracies of classes with small numbers of samples, we enrich the input of the 3D-ResNeXt spectral-spatial feature learning network by additional spectral feature learning, and finally use a loss function modified by label smoothing strategy to solve the imbalance of classes. The experimental results on three popular HSI datasets demonstrate the superiority of our proposed network and an effective improvement in the accuracies especially for the classes with small numbers of training samples.


2021 ◽  
Vol 2021 ◽  
pp. 1-16
Author(s):  
Xiong Zhao ◽  
Tao Zuo ◽  
Xinyu Hu

Most of the current visual Simultaneous Localization and Mapping (SLAM) algorithms are designed based on the assumption of a static environment, and their robustness and accuracy in the dynamic environment do not behave well. The reason is that moving objects in the scene will cause the mismatch of features in the pose estimation process, which further affects its positioning and mapping accuracy. In the meantime, the three-dimensional semantic map plays a key role in mobile robot navigation, path planning, and other tasks. In this paper, we present OFM-SLAM: Optical Flow combining MASK-RCNN SLAM, a novel visual SLAM for semantic mapping in dynamic indoor environments. Firstly, we use the Mask-RCNN network to detect potential moving objects which can generate masks of dynamic objects. Secondly, an optical flow method is adopted to detect dynamic feature points. Then, we combine the optical flow method and the MASK-RCNN for full dynamic points’ culling, and the SLAM system is able to track without these dynamic points. Finally, the semantic labels obtained from MASK-RCNN are mapped to the point cloud for generating a three-dimensional semantic map that only contains the static parts of the scenes and their semantic information. We evaluate our system in public TUM datasets. The results of our experiments demonstrate that our system is more effective in dynamic scenarios, and the OFM-SLAM can estimate the camera pose more accurately and acquire a more precise localization in the high dynamic environment.


Entropy ◽  
2021 ◽  
Vol 23 (10) ◽  
pp. 1316
Author(s):  
Kuiyong Song ◽  
Lianke Zhou ◽  
Hongbin Wang

Vigilance estimation of drivers is a hot research field of current traffic safety. Wearable devices can monitor information regarding the driver’s state in real time, which is then analyzed by a data analysis model to provide an estimation of vigilance. The accuracy of the data analysis model directly affects the effect of vigilance estimation. In this paper, we propose a deep coupling recurrent auto-encoder (DCRA) that combines electroencephalography (EEG) and electrooculography (EOG). This model uses a coupling layer to connect two single-modal auto-encoders to construct a joint objective loss function optimization model, which consists of single-modal loss and multi-modal loss. The single-modal loss is measured by Euclidean distance, and the multi-modal loss is measured by a Mahalanobis distance of metric learning, which can effectively reflect the distance between different modal data so that the distance between different modes can be described more accurately in the new feature space based on the metric matrix. In order to ensure gradient stability in the long sequence learning process, a multi-layer gated recurrent unit (GRU) auto-encoder model was adopted. The DCRA integrates data feature extraction and feature fusion. Relevant comparative experiments show that the DCRA is better than the single-modal method and the latest multi-modal fusion. The DCRA has a lower root mean square error (RMSE) and a higher Pearson correlation coefficient (PCC).


2021 ◽  
Vol 2021 ◽  
pp. 1-15
Author(s):  
Tao Liu ◽  
Jingxiong Wu ◽  
Jingfa Lei ◽  
Xue Wang ◽  
Bingqi Zhang ◽  
...  

In order to investigate the variation of three-dimensional metal surface topography during fatigue process, a three-dimensional (3D) topography acquisition platform was built with an in situ tensile tester and a three-dimensional profilometer. Q235 steel specimens were chosen as research objects, and the three-dimensional surface topography information at various stages of fatigue damage was obtained. Through the characterization of three-dimensional roughness, combined with surface height distribution and multifractal analysis, the variations of metal surface topography in the fatigue process were described. Results show that the arithmetic mean deviation of the surface (Sa), the width of the multifractal spectrum (Δα), and the mean value of surface height distribution (μ) and its standard deviation (δ) increase nonlinearly with the increase of fatigue cycles. The rate of fatigue damage is slow in the early stage and high in the middle and late stages. The surface height distribution amplitude (A) decreases with the increase of fatigue cycles, which indicates that the height data concentration decreases, and the metal surface becomes uneven. The Bayesian data fusion method was applied to establish a nonlinear mapping between the topography features and the damage, with the above five characteristic parameters (Sa, Δα, A, μ, and δ) as the data layer. Finally, a surface topography feature fusion method is proposed, and a case study is conducted to verify its applicability. The research results can provide reference for fatigue damage assessment.


Sign in / Sign up

Export Citation Format

Share Document