scholarly journals Hierarchical-Matching-Based Online and Real-Time Multi-Object Tracking with Deep Appearance Features

Algorithms ◽  
2020 ◽  
Vol 13 (4) ◽  
pp. 80 ◽  
Author(s):  
Qingge Ji ◽  
Haoqiang Yu ◽  
Xiao Wu

Based on tracking-by-detection, we propose a hierarchical-matching-based online and real-time multi-object tracking approach with deep appearance features, which can effectively reduce the false positives (FP) in tracking. For the purpose of increasing the accuracy rate of data association, we define the trajectory confidence using its position information, appearance information, and the information of historical relevant detections, after which we can classify the trajectories into different levels. In order to obtain discriminative appearance features, we developed a deep convolutional neural network to extract the appearance features of objects and trained it on a large-scale pedestrian re-identification dataset. Last but not least, we used the proposed diverse and hierarchical matching strategy to associate detection and trajectory sets. Experimental results on the MOT benchmark dataset show that our proposed approach performs well against other online methods, especially for the metrics of FP and frames per second (FPS).

Author(s):  
Amal Bouti ◽  
Mohamed Adnane Mahraz ◽  
Jamal Riffi ◽  
Hamid Tairi

In this chapter, the authors report a system for detection and classification of road signs. This system consists of two parts. The first part detects the road signs in real time. The second part classifies the German traffic signs (GTSRB) dataset and makes the prediction using the road signs detected in the first part to test the effectiveness. The authors used HOG and SVM in the detection part to detect the road signs captured by the camera. Then they used a convolutional neural network based on the LeNet model in which some modifications were added in the classification part. The system obtains an accuracy rate of 96.85% in the detection part and 96.23% in the classification part.


2021 ◽  
Vol 10 (11) ◽  
pp. 25420-25430
Author(s):  
Sofiane HADJI

Modeling large-scale flood inundation requires weeks of calculations using complex fluid software. The state-of-the-art in operational hydraulic modeling does not currently allow flood real-time forecasting fields. Data driven models have small computational costs and fast computation times and may be useful to overcome this problem. In this paper, we propose a new modeling approach based on a coupled of Hydrodynamics finite element model and Multi-headed Deep convolutional neural network (MH-CNN) with rain precipitations as input to forecast rapidly the water depth reached in large floodplain with few hours-ahead. For this purpose, one first builds a database containing different simulations of the physical model according to several rain precipitation scenarios (historic and synthetic). The multi-headed convolutional neural network is then trained using the constructed database to predict water depths. The pre-trained model is applied successfully to simulate the real July 2014 flood inundation in an 870 km2 area of La Nive watershed in the south west of France. Because rain precipitation forecast data is more accessible than discharge one, this approach offers great potential for real-time flood modelling for ungauged large-scale territories, which represent a large part of floodplain in the world.


2020 ◽  
Vol 2020 (10) ◽  
pp. 181-1-181-7
Author(s):  
Takahiro Kudo ◽  
Takanori Fujisawa ◽  
Takuro Yamaguchi ◽  
Masaaki Ikehara

Image deconvolution has been an important issue recently. It has two kinds of approaches: non-blind and blind. Non-blind deconvolution is a classic problem of image deblurring, which assumes that the PSF is known and does not change universally in space. Recently, Convolutional Neural Network (CNN) has been used for non-blind deconvolution. Though CNNs can deal with complex changes for unknown images, some CNN-based conventional methods can only handle small PSFs and does not consider the use of large PSFs in the real world. In this paper we propose a non-blind deconvolution framework based on a CNN that can remove large scale ringing in a deblurred image. Our method has three key points. The first is that our network architecture is able to preserve both large and small features in the image. The second is that the training dataset is created to preserve the details. The third is that we extend the images to minimize the effects of large ringing on the image borders. In our experiments, we used three kinds of large PSFs and were able to observe high-precision results from our method both quantitatively and qualitatively.


Sensors ◽  
2021 ◽  
Vol 21 (14) ◽  
pp. 4916
Author(s):  
Ali Usman Gondal ◽  
Muhammad Imran Sadiq ◽  
Tariq Ali ◽  
Muhammad Irfan ◽  
Ahmad Shaf ◽  
...  

Urbanization is a big concern for both developed and developing countries in recent years. People shift themselves and their families to urban areas for the sake of better education and a modern lifestyle. Due to rapid urbanization, cities are facing huge challenges, one of which is waste management, as the volume of waste is directly proportional to the people living in the city. The municipalities and the city administrations use the traditional wastage classification techniques which are manual, very slow, inefficient and costly. Therefore, automatic waste classification and management is essential for the cities that are being urbanized for the better recycling of waste. Better recycling of waste gives the opportunity to reduce the amount of waste sent to landfills by reducing the need to collect new raw material. In this paper, the idea of a real-time smart waste classification model is presented that uses a hybrid approach to classify waste into various classes. Two machine learning models, a multilayer perceptron and multilayer convolutional neural network (ML-CNN), are implemented. The multilayer perceptron is used to provide binary classification, i.e., metal or non-metal waste, and the CNN identifies the class of non-metal waste. A camera is placed in front of the waste conveyor belt, which takes a picture of the waste and classifies it. Upon successful classification, an automatic hand hammer is used to push the waste into the assigned labeled bucket. Experiments were carried out in a real-time environment with image segmentation. The training, testing, and validation accuracy of the purposed model was 0.99% under different training batches with different input features.


Sign in / Sign up

Export Citation Format

Share Document