Machine Vision and Applications
Latest Publications


TOTAL DOCUMENTS

1780
(FIVE YEARS 268)

H-INDEX

64
(FIVE YEARS 6)

Published By Springer-Verlag

1432-1769, 0932-8092

2022 ◽  
Vol 33 (1) ◽  
Author(s):  
Watcharin Sarachai ◽  
Jakramate Bootkrajang ◽  
Jeerayut Chaijaruwanich ◽  
Samerkae Somhom

2022 ◽  
Vol 33 (1) ◽  
Author(s):  
Federico Vaccaro ◽  
Marco Bertini ◽  
Tiberio Uricchio ◽  
Alberto Del Bimbo
Keyword(s):  

2021 ◽  
Vol 33 (1) ◽  
Author(s):  
David Freire-Obregón ◽  
Paola Barra ◽  
Modesto Castrillón-Santana ◽  
Maria De Marsico

AbstractAccording to the Wall Street Journal, one billion surveillance cameras will be deployed around the world by 2021. This amount of information can be hardly managed by humans. Using a Inflated 3D ConvNet as backbone, this paper introduces a novel automatic violence detection approach that outperforms state-of-the-art existing proposals. Most of those proposals consider a pre-processing step to only focus on some regions of interest in the scene, i.e., those actually containing a human subject. In this regard, this paper also reports the results of an extensive analysis on whether and how the context can affect or not the adopted classifier performance. The experiments show that context-free footage yields substantial deterioration of the classifier performance (2% to 5%) on publicly available datasets. However, they also demonstrate that performance stabilizes in context-free settings, no matter the level of context restriction applied. Finally, a cross-dataset experiment investigates the generalizability of results obtained in a single-collection experiment (same dataset used for training and testing) to cross-collection settings (different datasets used for training and testing).


2021 ◽  
Vol 33 (1) ◽  
Author(s):  
Ning Yang ◽  
Yunlong Han ◽  
Jun Fang ◽  
Weijun Zhong ◽  
Anlin Xu

2021 ◽  
Vol 33 (1) ◽  
Author(s):  
Mariana-Iuliana Georgescu ◽  
Georgian-Emilian Duţǎ ◽  
Radu Tudor Ionescu

2021 ◽  
Vol 33 (1) ◽  
Author(s):  
Kuldeep Purohit ◽  
Subeesh Vasu ◽  
M. Purnachandra Rao ◽  
A. N. Rajagopalan

2021 ◽  
Vol 33 (1) ◽  
Author(s):  
Majedaldein Almahasneh ◽  
Adeline Paiement ◽  
Xianghua Xie ◽  
Jean Aboudarham

AbstractPrecisely localising solar Active Regions (AR) from multi-spectral images is a challenging but important task in understanding solar activity and its influence on space weather. A main challenge comes from each modality capturing a different location of the 3D objects, as opposed to typical multi-spectral imaging scenarios where all image bands observe the same scene. Thus, we refer to this special multi-spectral scenario as multi-layer. We present a multi-task deep learning framework that exploits the dependencies between image bands to produce 3D AR localisation (segmentation and detection) where different image bands (and physical locations) have their own set of results. Furthermore, to address the difficulty of producing dense AR annotations for training supervised machine learning (ML) algorithms, we adapt a training strategy based on weak labels (i.e. bounding boxes) in a recursive manner. We compare our detection and segmentation stages against baseline approaches for solar image analysis (multi-channel coronal hole detection, SPOCA for ARs) and state-of-the-art deep learning methods (Faster RCNN, U-Net). Additionally, both detection and segmentation stages are quantitatively validated on artificially created data of similar spatial configurations made from annotated multi-modal magnetic resonance images. Our framework achieves an average of 0.72 IoU (segmentation) and 0.90 F1 score (detection) across all modalities, comparing to the best performing baseline methods with scores of 0.53 and 0.58, respectively, on the artificial dataset, and 0.84 F1 score in the AR detection task comparing to baseline of 0.82 F1 score. Our segmentation results are qualitatively validated by an expert on real ARs.


2021 ◽  
Vol 33 (1) ◽  
Author(s):  
Meiling Fang ◽  
Naser Damer ◽  
Fadi Boutros ◽  
Florian Kirchbuchner ◽  
Arjan Kuijper

AbstractIris Presentation Attack Detection (PAD) algorithms address the vulnerability of iris recognition systems to presentation attacks. With the great success of deep learning methods in various computer vision fields, neural network-based iris PAD algorithms emerged. However, most PAD networks suffer from overfitting due to insufficient iris data variability. Therefore, we explore the impact of various data augmentation techniques on performance and the generalizability of iris PAD. We apply several data augmentation methods to generate variability, such as shift, rotation, and brightness. We provide in-depth analyses of the overlapping effect of these methods on performance. In addition to these widely used augmentation techniques, we also propose an augmentation selection protocol based on the assumption that various augmentation techniques contribute differently to the PAD performance. Moreover, two fusion methods are performed for more comparisons: the strategy-level and the score-level combination. We demonstrate experiments on two fine-tuned models and one trained from the scratch network and perform on the datasets in the Iris-LivDet-2017 competition designed for generalizability evaluation. Our experimental results show that augmentation methods improve iris PAD performance in many cases. Our least overlap-based augmentation selection protocol achieves the lower error rates for two networks. Besides, the shift augmentation strategy also exceeds state-of-the-art (SoTA) algorithms on the Clarkson and IIITD-WVU datasets.


Sign in / Sign up

Export Citation Format

Share Document