A change-detection algorithm based on structure and colour

Author(s):  
L. Di Stefano ◽  
S. Mattoccia ◽  
M. Mola
2005 ◽  
Vol 53 (8) ◽  
pp. 2961-2974 ◽  
Author(s):  
F. Desobry ◽  
M. Davy ◽  
C. Doncarli

2011 ◽  
Vol 24 ◽  
pp. 252-256 ◽  
Author(s):  
Wei Cui ◽  
Zhenhong Jia ◽  
Xizhong Qin ◽  
Jie Yang ◽  
Yingjie Hu

Author(s):  
Gulnaz Alimjan ◽  
Yiliyaer Jiaermuhamaiti ◽  
Huxidan Jumahong ◽  
Shuangling Zhu ◽  
Pazilat Nurmamat

Various UNet architecture-based image change detection algorithms promote the development of image change detection, but there are still some defects. First, under the encoder–decoder framework, the low-level features are extracted many times in multiple dimensions, which generates redundant information; second, the relationship between each feature layer is not modeled so sufficiently that it cannot produce the optimal feature differentiation representation. This paper proposes a remote image change detection algorithm based on the multi-feature self-attention fusion mechanism UNet network, abbreviated as MFSAF UNet (multi-feature self-attention fusion UNet). We attempt to add multi-feature self-attention mechanism between the encoder and decoder of UNet to obtain richer context dependence and overcome the two above-mentioned restrictions. Since the capacity of convolution-based UNet network is directly proportional to network depth, and a deeper convolutional network means more training parameters, so the convolution of each layer of UNet is replaced as a separated convolution, which makes the entire network to be lighter and the model’s execution efficiency is slightly better than the traditional convolution operation. In addition to these, another innovation point of this paper is using preference to control loss function and meet the demands for different accuracies and recall rates. The simulation test results verify the validity and robustness of this approach.


Author(s):  
A. Schlichting ◽  
C. Brenner

LiDAR sensors are proven sensors for accurate vehicle localization. Instead of detecting and matching features in the LiDAR data, we want to use the entire information provided by the scanners. As dynamic objects, like cars, pedestrians or even construction sites could lead to wrong localization results, we use a change detection algorithm to detect these objects in the reference data. If an object occurs in a certain number of measurements at the same position, we mark it and every containing point as static. In the next step, we merge the data of the single measurement epochs to one reference dataset, whereby we only use static points. Further, we also use a classification algorithm to detect trees. <br><br> For the online localization of the vehicle, we use simulated data of a vertical aligned automotive LiDAR sensor. As we only want to use static objects in this case as well, we use a random forest classifier to detect dynamic scan points online. Since the automotive data is derived from the LiDAR Mobile Mapping System, we are able to use the labelled objects from the reference data generation step to create the training data and further to detect dynamic objects online. The localization then can be done by a point to image correlation method using only static objects. We achieved a localization standard deviation of about 5 cm (position) and 0.06° (heading), and were able to successfully localize the vehicle in about 93 % of the cases along a trajectory of 13 km in Hannover, Germany.


2015 ◽  
Vol 11 (2) ◽  
pp. 160 ◽  
Author(s):  
Abdullah Baqasah ◽  
Eric Pardede ◽  
Wenny Rahayu ◽  
Irena Holubova

Sign in / Sign up

Export Citation Format

Share Document