prediction module
Recently Published Documents


TOTAL DOCUMENTS

96
(FIVE YEARS 43)

H-INDEX

9
(FIVE YEARS 3)

2022 ◽  
Vol 2022 ◽  
pp. 1-14
Author(s):  
Liangliang Duan

Deep encoder-decoder networks have been adopted for saliency detection and achieved state-of-the-art performance. However, most existing saliency models usually fail to detect very small salient objects. In this paper, we propose a multitask architecture, M2Net, and a novel centerness-aware loss for salient object detection. The proposed M2Net aims to solve saliency prediction and centerness prediction simultaneously. Specifically, the network architecture is composed of a bottom-up encoder module, top-down decoder module, and centerness prediction module. In addition, different from binary cross entropy, the proposed centerness-aware loss can guide the proposed M2Net to uniformly highlight the entire salient regions with well-defined object boundaries. Experimental results on five benchmark saliency datasets demonstrate that M2Net outperforms state-of-the-art methods on different evaluation metrics.


Machines ◽  
2021 ◽  
Vol 10 (1) ◽  
pp. 19
Author(s):  
Mu Chen ◽  
Huaici Zhao ◽  
Pengfei Liu

Three-dimensional (3D) object detection is an important task in the field of machine vision, in which the detection of 3D objects using monocular vision is even more challenging. We observe that most of the existing monocular methods focus on the design of the feature extraction framework or embedded geometric constraints, but ignore the possible errors in the intermediate process of the detection pipeline. These errors may be further amplified in the subsequent processes. After exploring the existing detection framework of keypoints, we find that the accuracy of keypoints prediction will seriously affect the solution of 3D object position. Therefore, we propose a novel keypoints uncertainty prediction network (KUP-Net) for monocular 3D object detection. In this work, we design an uncertainty prediction module to characterize the uncertainty that exists in keypoint prediction. Then, the uncertainty is used for joint optimization with object position. In addition, we adopt position-encoding to assist the uncertainty prediction, and use a timing coefficient to optimize the learning process. The experiments on our detector are conducted on the KITTI benchmark. For the two levels of easy and moderate, we achieve accuracy of 17.26 and 11.78 in AP3D, and achieve accuracy of 23.59 and 16.63 in APBEV, which are higher than the latest method KM3D.


2021 ◽  
Vol 13 (23) ◽  
pp. 4941
Author(s):  
Rukhshanda Hussain ◽  
Yash Karbhari ◽  
Muhammad Fazal Ijaz ◽  
Marcin Woźniak ◽  
Pawan Kumar Singh ◽  
...  

Recently, deep learning-based methods, especially utilizing fully convolutional neural networks, have shown extraordinary performance in salient object detection. Despite its success, the clean boundary detection of the saliency objects is still a challenging task. Most of the contemporary methods focus on exclusive edge detection modules in order to avoid noisy boundaries. In this work, we propose leveraging on the extraction of finer semantic features from multiple encoding layers and attentively re-utilize it in the generation of the final segmentation result. The proposed Revise-Net model is divided into three parts: (a) the prediction module, (b) a residual enhancement module, and (c) reverse attention modules. Firstly, we generate the coarse saliency map through the prediction modules, which are fine-tuned in the enhancement module. Finally, multiple reverse attention modules at varying scales are cascaded between the two networks to guide the prediction module by employing the intermediate segmentation maps generated at each downsampling level of the REM. Our method efficiently classifies the boundary pixels using a combination of binary cross-entropy, similarity index, and intersection over union losses at the pixel, patch, and map levels, thereby effectively segmenting the saliency objects in an image. In comparison with several state-of-the-art frameworks, our proposed Revise-Net model outperforms them with a significant margin on three publicly available datasets, DUTS-TE, ECSSD, and HKU-IS, both on regional and boundary estimation measures.


2021 ◽  
Vol 87 (12) ◽  
pp. 1008-1012
Author(s):  
Keisuke GOTO ◽  
Hiroaki AIZAWA ◽  
Kunihito KATO ◽  
Yoshihiro HARADA ◽  
Minori NOGUCHI ◽  
...  

Sensors ◽  
2021 ◽  
Vol 21 (14) ◽  
pp. 4701
Author(s):  
Yunxia Liu ◽  
Zeyu Zou ◽  
Yang Yang ◽  
Ngai-Fong Bonnie Law ◽  
Anil Anthony Bharath

Source camera identification has long been a hot topic in the field of image forensics. Besides conventional feature engineering algorithms developed based on studying the traces left upon shooting, several deep-learning-based methods have also emerged recently. However, identification performance is susceptible to image content and is far from satisfactory for small image patches in real demanding applications. In this paper, an efficient patch-level source camera identification method is proposed based on a convolutional neural network. First, in order to obtain improved robustness with reduced training cost, representative patches are selected according to multiple criteria for enhanced diversity in training data. Second, a fine-grained multiscale deep residual prediction module is proposed to reduce the impact of scene content. Finally, a modified VGG network is proposed for source camera identification at brand, model, and instance levels. A more critical patch-level evaluation protocol is also proposed for fair performance comparison. Abundant experimental results show that the proposed method achieves better results as compared with the state-of-the-art algorithms.


Author(s):  
Sadaf Saqib ◽  

The Internet of Things (IoT) has achieved an upset in a considerable lot of the circles of our current lives, like automobile, medical services offices, home automation, retail, education, manufacturing, and many more. The Agriculture and Farming ventures significantly affect the acquaintance of the IoT with the world. Machine learning (ML) is a part of artificial intelligence (AI) that permits software applications to turn out to be more precise at foreseeing results without being expressly customized to do as such. It uses historical data as input to predict new result values. In the event, a specific industry has sufficient recorded information to help the machine "learn", AI or ML can create outstanding outcomes. Farming is likewise one such important industry profiting and advancing from machine learning at large. ML can possibly add to the total lifecycle of farming, at all phases. This incorporates computer vision, automated irrigation, and harvesting, predicting the soil, weather, temperature, moisture values, and robots for picking off the crude harvest. In this paper, I'll work on a smart agricultural information monitoring framework that gathers the necessary information from the IoT sensors set in the field, measures it, and drives it, from where it streams to store in the cloud space. The information is then shipped off the prediction module where the necessary analysis is done using ML algorithms and afterward sent to the UI for its corresponding application.


Robotica ◽  
2021 ◽  
pp. 1-19
Author(s):  
H. S. Hewawasam ◽  
M. Yousef Ibrahim ◽  
Gayan Kahandawa ◽  
T. A. Choudhury

Abstract This paper presents a new algorithm to navigate robots in dynamically cluttered environments. The proposed algorithm uses basic concepts of space attraction (hence the term Agoraphilic) to navigate robots through dynamic obstacles. The new algorithm in this paper is an advanced development of the original Agoraphilic navigation algorithm that was only able to navigate robots in static environments. The Agoraphilic algorithm does not look for obstacles (problems) to avoid but rather for a free space (solutions) to follow. Therefore, it is also described as an optimistic navigation algorithm. This algorithm uses only one attractive force created by the available free space. The free-space concept allows the Agoraphilic algorithm to overcome inherited challenges of general navigation algorithms. However, the original Agoraphilic algorithm has the limitation in navigating robots only in static, not in dynamic environments. The presented algorithm was developed to address this limitation of the original Agoraphilic algorithm. The new algorithm uses a developed object tracking module to identify the time-varying free spaces by tracking moving obstacles. The capacity of the algorithm was further strengthened by the new prediction module. Future space prediction allowed the algorithm to make decisions considering future growing/diminishing free spaces. This paper also includes a bench-marking study of the new algorithm compared with a recently published APF-based algorithm under a similar operating environment. Furthermore, the algorithm was validated based on experimental tests and simulation tests.


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Li Cen Lim ◽  
Yee Ying Lim ◽  
Yee Siew Choong

Abstract B-cell epitope will be recognized and attached to the surface of receptors in B-lymphocytes to trigger immune response, thus are the vital elements in the field of epitope-based vaccine design, antibody production and therapeutic development. However, the experimental approaches in mapping epitopes are time consuming and costly. Computational prediction could offer an unbiased preliminary selection to reduce the number of epitopes for experimental validation. The deposited B-cell epitopes in the databases are those with experimentally determined positive/negative peptides and some are ambiguous resulted from different experimental methods. Prior to the development of B-cell epitope prediction module, the available dataset need to be handled with care. In this work, we first pre-processed the B-cell epitope dataset prior to B-cell epitopes prediction based on pattern recognition using support vector machine (SVM). By using only the absolute epitopes and non-epitopes, the datasets were classified into five categories of pathogen and worked on the 6-mers peptide sequences. The pre-processing of the datasets have improved the B-cell epitope prediction performance up to 99.1 % accuracy and showed significant improvement in cross validation results. It could be useful when incorporated with physicochemical propensity ranking in the future for the development of B-cell epitope prediction module.


2021 ◽  
Vol 2021 ◽  
pp. 1-14
Author(s):  
Hongying Zheng ◽  
Zhiqiang Zhou ◽  
Jianyong Chen

An accurate prediction of stock market index is important for investors to reduce financial risk. Although quite a number of deep learning methods have been developed for the stock prediction, some fundamental problems, such as weak generalization ability and overfitting in training, need to be solved. In this paper, a new deep learning model named Random Long Short-Term Memory (RLSTM) is proposed to get a better predicting result. RLSTM includes prediction module, prevention module, and three full connection layers. Input of the prediction module is a stock or an index which needs to be predicted. That of the prevention module is a random number series. With the index of Shanghai Securities Composite Index (SSEC) and Standard & Poor’s 500 (S&P500), simulations show that the proposed RLSTM can mitigate the overfitting and outperform others in accuracy of prediction.


Author(s):  
Jinbo Xing ◽  
Wenbo Hu ◽  
Yuechen Zhang ◽  
Tien-Tsin Wong

AbstractA popular and challenging task in video research, frame interpolation aims to increase the frame rate of video. Most existing methods employ a fixed motion model, e.g., linear, quadratic, or cubic, to estimate the intermediate warping field. However, such fixed motion models cannot well represent the complicated non-linear motions in the real world or rendered animations. Instead, we present an adaptive flow prediction module to better approximate the complex motions in video. Furthermore, interpolating just one intermediate frame between consecutive input frames may be insufficient for complicated non-linear motions. To enable multi-frame interpolation, we introduce the time as a control variable when interpolating frames between original ones in our generic adaptive flow prediction module. Qualitative and quantitative experimental results show that our method can produce high-quality results and outperforms the existing state-of-the-art methods on popular public datasets.


Sign in / Sign up

Export Citation Format

Share Document