scholarly journals Detecting Extratropical Cyclones of the Northern Hemisphere with Single Shot Detector

2022 ◽  
Vol 14 (2) ◽  
pp. 254
Author(s):  
Minjing Shi ◽  
Pengfei He ◽  
Yuli Shi

In this paper, we propose a deep learning-based model to detect extratropical cyclones (ETCs) of the northern hemisphere, while developing a novel workflow of processing images and generating labels for ETCs. We first labeled the cyclone center by adapting an approach from Bonfanti et al. in 2017 and set up criteria of labeling ETCs of three categories: developing, mature, and declining stages. We then gave a framework of labeling and preprocessing the images in our dataset. Once the images and labels were ready to serve as inputs, an object detection model was built with Single Shot Detector (SSD) and adjusted to fit the format of the dataset. We trained and evaluated our model with our labeled dataset on two settings (binary and multiclass classifications), while keeping a record of the results. We found that the model achieves relatively high performance with detecting ETCs of mature stage (mean Average Precision is 86.64%), and an acceptable result for detecting ETCs of all three categories (mean Average Precision 79.34%). The single-shot detector model can succeed in detecting ETCs of different stages, and it has demonstrated great potential in the future applications of ETC detection in other relevant settings.

2021 ◽  
Author(s):  
Komuravelli Prashanth ◽  
Kalidas Yeturu

<div>There are millions of scanned documents worldwide in around 4 thousand languages. Searching for information in a scanned document requires a text layer to be available and indexed. Preparation of a text layer requires recognition of character and sub-region patterns and associating with a human interpretation. Developing an optical character recognition (OCR) system for each and every language is a very difficult task if not impossible. There is a strong need for systems that add on top of the existing OCR technologies by learning from them and unifying disparate multitude of many a system. In this regard, we propose an algorithm that leverages the fact that we are dealing with scanned documents of handwritten text regions from across diverse domains and language settings. We observe that the text regions have consistent bounding box sizes and any large font or tiny font scenarios can be handled in preprocessing or postprocessing phases. The image subregions are smaller in size in scanned text documents compared to subregions formed by common objects in general purpose images. We propose and validate the hypothesis that a much simpler convolution neural network (CNN) having very few layers and less number of filters can be used for detecting individual subregion classes. For detection of several hundreds of classes, multiple such simpler models can be pooled to operate simultaneously on a document. The advantage of going by pools of subregion specific models is the ability to deal with incremental addition of hundreds of newer classes over time, without disturbing the previous models in the continual learning scenario. Such an approach has distinctive advantage over using a single monolithic model where subregions classes share and interfere via a bulky common neural network. We report here an efficient algorithm for building a subregion specific lightweight CNN models. The training data for the CNN proposed, requires engineering synthetic data points that consider both pattern of interest and non-patterns as well. We propose and validate the hypothesis that an image canvas in which optimal amount of pattern and non-pattern can be formulated using a means squared error loss function to influence filter for training from the data. The CNN hence trained has the capability to identify the character-object in presence of several other objects on a generalized test image of a scanned document. In this setting some of the key observations are in a CNN, learning a filter depends not only on the abundance of patterns of interest but also on the presence of a non-pattern context. Our experiments have led to some of the key observations - (i) a pattern cannot be over-expressed in isolation, (ii) a pattern cannot be under-xpressed as well, (iii) a non-pattern can be of salt and pepper type noise and finally (iv) it is sufficient to provide a non-pattern context to a modest representation of a pattern to result in strong individual sub-region class models. We have carried out studies and reported \textit{mean average precision} scores on various data sets including (1) MNIST digits(95.77), (2) E-MNIST capital alphabet(81.26), (3) EMNIST small alphabet(73.32) (4) Kannada digits(95.77), (5) Kannada letters(90.34), (6) Devanagari letters(100) (7) Telugu words(93.20) (8) Devanagari words(93.20) and also on medical prescriptions and observed high-performance metrics of mean average precision over 90%. The algorithm serves as a kernel in the automatic annotation of digital documents in diverse scenarios such as annotation of ancient manuscripts and hand-written health records.</div>


Information ◽  
2019 ◽  
Vol 10 (6) ◽  
pp. 209 ◽  
Author(s):  
Abdul Jabbar Saleh ◽  
Asif Karim ◽  
Bharanidharan Shanmugam ◽  
Sami Azam ◽  
Krishnan Kannoorpatti ◽  
...  

Spam emails, also known as non-self, are unsolicited commercial or malicious emails, sent to affect either a single individual or a corporation or a group of people. Besides advertising, these may contain links to phishing or malware hosting websites set up to steal confidential information. In this paper, a study of the effectiveness of using a Negative Selection Algorithm (NSA) for anomaly detection applied to spam filtering is presented. NSA has a high performance and a low false detection rate. The designed framework intelligently works through three detection phases to finally determine an email’s legitimacy based on the knowledge gathered in the training phase. The system operates by elimination through Negative Selection similar to the functionality of T-cells’ in biological systems. It has been observed that with the inclusion of more datasets, the performance continues to improve, resulting in a 6% increase of True Positive and True Negative detection rate while achieving an actual detection rate of spam and ham of 98.5%. The model has been further compared against similar studies, and the result shows that the proposed system results in an increase of 2 to 15% in the correct detection rate of spam and ham.


2012 ◽  
Vol 69 (2) ◽  
pp. 725-740 ◽  
Author(s):  
Jana Čampa ◽  
Heini Wernli

Abstract Development of extratropical cyclones can be seen as an interplay of three positive potential vorticity anomalies: an upper-level stratospheric intrusion, low-tropospheric diabatically produced potential vorticity (PV), and a warm anomaly at the surface acting as a surrogate PV anomaly. This study, based on the interim ECMWF Re-Analysis (ERA-Interim) dataset, quantifies the amplitude of the PV anomalies of mature extratropical cyclones in different regions in the Northern Hemisphere on a climatological basis. A tracking algorithm is applied to sea level pressure (SLP) fields to identify cyclone tracks. Surface potential temperature anomalies Δθ and vertical profiles of PV anomalies ΔPV are calculated at the time of the cyclones’ minimum SLP in a vertical cylinder around the surface cyclone center. To compare the cyclones’ characteristics they are grouped according to their location and intensity. Composite ΔPV profiles are calculated for each region and intensity class at the time of minimum SLP and during the cyclone intensification phase. In the mature stage all three anomalies are on average larger for intense than for weak winter cyclones [e.g., 0.6 versus 0.2 potential vorticity units (PVU; 1 PVU = 10−6 K kg−1 m2 s−1) at lower levels, and 1.5 versus 0.5 PVU at upper levels]. The regional variability of the cyclones’ vertical structure and the profile evolution is prominent (cyclones in some regions are more sensitive to the amplitude of a particular anomaly than in other regions). Values of Δθ and low-level ΔPV are on average larger in the western parts of the oceans than in the eastern parts. Results for summer are qualitatively similar, except for distinctively weaker surface Δθ values.


Plants ◽  
2020 ◽  
Vol 9 (11) ◽  
pp. 1451
Author(s):  
Muhammad Hammad Saleem ◽  
Sapna Khanchi ◽  
Johan Potgieter ◽  
Khalid Mahmood Arif

The identification of plant disease is an imperative part of crop monitoring systems. Computer vision and deep learning (DL) techniques have been proven to be state-of-the-art to address various agricultural problems. This research performed the complex tasks of localization and classification of the disease in plant leaves. In this regard, three DL meta-architectures including the Single Shot MultiBox Detector (SSD), Faster Region-based Convolutional Neural Network (RCNN), and Region-based Fully Convolutional Networks (RFCN) were applied by using the TensorFlow object detection framework. All the DL models were trained/tested on a controlled environment dataset to recognize the disease in plant species. Moreover, an improvement in the mean average precision of the best-obtained deep learning architecture was attempted through different state-of-the-art deep learning optimizers. The SSD model trained with an Adam optimizer exhibited the highest mean average precision (mAP) of 73.07%. The successful identification of 26 different types of defected and 12 types of healthy leaves in a single framework proved the novelty of the work. In the future, the proposed detection methodology can also be adopted for other agricultural applications. Moreover, the generated weights can be reused for future real-time detection of plant disease in a controlled/uncontrolled environment.


Diagnostics ◽  
2020 ◽  
Vol 10 (12) ◽  
pp. 1113
Author(s):  
Djeane Debora Onthoni ◽  
Ting-Wen Sheng ◽  
Prasan Kumar Sahoo ◽  
Li-Jen Wang ◽  
Pushpanjali Gupta

Total Kidney Volume (TKV) is essential for analyzing the progressive loss of renal function in Autosomal Dominant Polycystic Kidney Disease (ADPKD). Conventionally, to measure TKV from medical images, a radiologist needs to localize and segment the kidneys by defining and delineating the kidney’s boundary slice by slice. However, kidney localization is a time-consuming and challenging task considering the unstructured medical images from big data such as Contrast-enhanced Computed Tomography (CCT). This study aimed to design an automatic localization model of ADPKD using Artificial Intelligence. A robust detection model using CCT images, image preprocessing, and Single Shot Detector (SSD) Inception V2 Deep Learning (DL) model is designed here. The model is trained and evaluated with 110 CCT images that comprise 10,078 slices. The experimental results showed that our derived detection model outperformed other DL detectors in terms of Average Precision (AP) and mean Average Precision (mAP). We achieved mAP = 94% for image-wise testing and mAP = 82% for subject-wise testing, when threshold on Intersection over Union (IoU) = 0.5. This study proves that our derived automatic detection model can assist radiologist in locating and classifying the ADPKD kidneys precisely and rapidly in order to improve the segmentation task and TKV calculation.


2021 ◽  
Vol 8 (2) ◽  
pp. 357
Author(s):  
Muhammad Fadhlan Supriadi ◽  
Ema Rachmawati ◽  
Anditya Arifianto

<p class="Abstrak">Penggunaan ponsel sudah sangat erat dengan kehidupaan anak usia dini sehingga menimbulkan beberapa dampak negatif bagi anak usia dini terutama berkurangnya interaksi dengan dunia sekitarnya. Salah satu teknologi yang dapat dikembangkan pada ponsel adalah <em>computer vision. </em>Salah satu penggunaan <em>computer vision </em>adalah <em>object recognition</em> yang memberikan solusi untuk membantu mengenali objek.<em> </em>Pada penelitian ini dibangun sistem pengenalan objek benda di dalam rumah yang diaplikasikan pada ponsel yang diharapkan membantu anak usia dini mengenali benda disekitarnya. <em>MobileNet </em>merupakan salah satu <em>feature extraction</em> yang memiliki kinerja yang baik dan ringan digunakan pada perangkat ponsel. Arsitektur <em>MobileNet </em>terdiri dari <em>layer depthwise convolution </em>dan <em>layer pointwise </em><em>convolution </em>dalam mengekstraksi fitur<em>. </em>Percobaan ini juga menggunakan arsitektur <em>Single Shot Multibox Detector (SSD) </em>sebagai metode dalam mendeteksi objek<em>. Pre-trained model </em>dari dataset <em>COCO </em>digunakan pada eksperimen<em>,</em> untuk mengenali 20 jenis objek benda di dalam rumah. Dari hasil eksperimen, <em>MobileNetV2 </em>menghasilkan nilai <em>mean Average Precision (mAP)</em> yang lebih baik dibandingkan dengan <em>MobileNetV1 </em>dan<em> InceptionV2, </em>yaitu<em> </em>sebesar 99,34%.</p><p class="Abstrak"> </p><p class="Judul2"><strong><em>Abstract</em></strong></p><p class="Judul2"> <em>Mobile phone usage has been very close to early childhood life, so giving rise to some negative impact on early childhood, especially reduced interaction with the surrounding world. One of the technologies that can be developed on the cellphone is computer vision. One of the uses of computer vision is object recognition that provides solutions to help to recognize objects. This research builds a system for recognition objects inside in house that is developed on a cellphone that is expected to help early childhood recognize objects in the surrounding. MobileNet is one of feature extraction that has good performance and efficient use on a cellphone. MobileNet architecture consists of a depthwise convolution layer and pointwise convolution layer in extracting features. The experiment also uses the architecture of Single Shot Multibox Detector (SSD) as a method of detecting objects. We used MobileNet architecture as a pre-trained model that had previously been trained on COCO datasets, and implement transfer learning for 20 types of objects commonly found inside the house. The experimental result indicates that the mean Average Precision (mAP) of MobileNetV2 could exceed MobileNetV1 and InceptionV2 of 99.34%.</em></p>


2021 ◽  
Author(s):  
Komuravelli Prashanth ◽  
Kalidas Yeturu

<div>There are millions of scanned documents worldwide in around 4 thousand languages. Searching for information in a scanned document requires a text layer to be available and indexed. Preparation of a text layer requires recognition of character and sub-region patterns and associating with a human interpretation. Developing an optical character recognition (OCR) system for each and every language is a very difficult task if not impossible. There is a strong need for systems that add on top of the existing OCR technologies by learning from them and unifying disparate multitude of many a system. In this regard, we propose an algorithm that leverages the fact that we are dealing with scanned documents of handwritten text regions from across diverse domains and language settings. We observe that the text regions have consistent bounding box sizes and any large font or tiny font scenarios can be handled in preprocessing or postprocessing phases. The image subregions are smaller in size in scanned text documents compared to subregions formed by common objects in general purpose images. We propose and validate the hypothesis that a much simpler convolution neural network (CNN) having very few layers and less number of filters can be used for detecting individual subregion classes. For detection of several hundreds of classes, multiple such simpler models can be pooled to operate simultaneously on a document. The advantage of going by pools of subregion specific models is the ability to deal with incremental addition of hundreds of newer classes over time, without disturbing the previous models in the continual learning scenario. Such an approach has distinctive advantage over using a single monolithic model where subregions classes share and interfere via a bulky common neural network. We report here an efficient algorithm for building a subregion specific lightweight CNN models. The training data for the CNN proposed, requires engineering synthetic data points that consider both pattern of interest and non-patterns as well. We propose and validate the hypothesis that an image canvas in which optimal amount of pattern and non-pattern can be formulated using a means squared error loss function to influence filter for training from the data. The CNN hence trained has the capability to identify the character-object in presence of several other objects on a generalized test image of a scanned document. In this setting some of the key observations are in a CNN, learning a filter depends not only on the abundance of patterns of interest but also on the presence of a non-pattern context. Our experiments have led to some of the key observations - (i) a pattern cannot be over-expressed in isolation, (ii) a pattern cannot be under-xpressed as well, (iii) a non-pattern can be of salt and pepper type noise and finally (iv) it is sufficient to provide a non-pattern context to a modest representation of a pattern to result in strong individual sub-region class models. We have carried out studies and reported \textit{mean average precision} scores on various data sets including (1) MNIST digits(95.77), (2) E-MNIST capital alphabet(81.26), (3) EMNIST small alphabet(73.32) (4) Kannada digits(95.77), (5) Kannada letters(90.34), (6) Devanagari letters(100) (7) Telugu words(93.20) (8) Devanagari words(93.20) and also on medical prescriptions and observed high-performance metrics of mean average precision over 90%. The algorithm serves as a kernel in the automatic annotation of digital documents in diverse scenarios such as annotation of ancient manuscripts and hand-written health records.</div>


2021 ◽  
Author(s):  
Najmath Ottakath ◽  
Abdulla Al-Ali ◽  
Somaya Al Maadeed

Vehicles are a common sight on the road. Tracking and monitoring suspicious vehicles for identification due to high similarity in structure and form leads to difficulties in differentiating between them. The unique identity of a vehicle, the license plate is used here for this purpose. License plate detection is considered as an object detection task. Transfer learning on pre-trained state of art object detection models is an approach, which can perform this with better accuracy in terms of mean average precision. However, setting the right hyper-parameters needs multiple experiments. In this research, an evolutionary algorithm, genetic algorithm is used, which can optimize the hyper-parameters to achieve the best accuracy for the object detection model, YOLOv5. Further, the license plate was identified using OCR. This study concluded that hyper-parameter tuning achieved high accuracy in terms of mean average precision, achieving 98.25%, compared to 80% in initial parameter set providing an automated optimization. This license plate detected can be stored in a secure location and retrieved for re-identification. A decentralized storage or a secure cloud can be used to store the license plate. The application of this is most relevant to surveillance in high security locations where suspicious vehicles must be tracked.


2021 ◽  
Author(s):  
Komuravelli Prashanth ◽  
Kalidas Yeturu

<div>There are millions of scanned documents worldwide in around 4 thousand languages. Searching for information in a scanned document requires a text layer to be available and indexed. Preparation of a text layer requires recognition of character and sub-region patterns and associating with a human interpretation. Developing an optical character recognition (OCR) system for each and every language is a very difficult task if not impossible. There is a strong need for systems that add on top of the existing OCR technologies by learning from them and unifying disparate multitude of many a system. In this regard, we propose an algorithm that leverages the fact that we are dealing with scanned documents of handwritten text regions from across diverse domains and language settings. We observe that the text regions have consistent bounding box sizes and any large font or tiny font scenarios can be handled in preprocessing or postprocessing phases. The image subregions are smaller in size in scanned text documents compared to subregions formed by common objects in general purpose images. We propose and validate the hypothesis that a much simpler convolution neural network (CNN) having very few layers and less number of filters can be used for detecting individual subregion classes. For detection of several hundreds of classes, multiple such simpler models can be pooled to operate simultaneously on a document. The advantage of going by pools of subregion specific models is the ability to deal with incremental addition of hundreds of newer classes over time, without disturbing the previous models in the continual learning scenario. Such an approach has distinctive advantage over using a single monolithic model where subregions classes share and interfere via a bulky common neural network. We report here an efficient algorithm for building a subregion specific lightweight CNN models. The training data for the CNN proposed, requires engineering synthetic data points that consider both pattern of interest and non-patterns as well. We propose and validate the hypothesis that an image canvas in which optimal amount of pattern and non-pattern can be formulated using a means squared error loss function to influence filter for training from the data. The CNN hence trained has the capability to identify the character-object in presence of several other objects on a generalized test image of a scanned document. In this setting some of the key observations are in a CNN, learning a filter depends not only on the abundance of patterns of interest but also on the presence of a non-pattern context. Our experiments have led to some of the key observations - (i) a pattern cannot be over-expressed in isolation, (ii) a pattern cannot be under-xpressed as well, (iii) a non-pattern can be of salt and pepper type noise and finally (iv) it is sufficient to provide a non-pattern context to a modest representation of a pattern to result in strong individual sub-region class models. We have carried out studies and reported \textit{mean average precision} scores on various data sets including (1) MNIST digits(95.77), (2) E-MNIST capital alphabet(81.26), (3) EMNIST small alphabet(73.32) (4) Kannada digits(95.77), (5) Kannada letters(90.34), (6) Devanagari letters(100) (7) Telugu words(93.20) (8) Devanagari words(93.20) and also on medical prescriptions and observed high-performance metrics of mean average precision over 90%. The algorithm serves as a kernel in the automatic annotation of digital documents in diverse scenarios such as annotation of ancient manuscripts and hand-written health records.</div>


Author(s):  
Anthony Anggrawan ◽  
Azhari

Information searching based on users’ query, which is hopefully able to find the documents based on users’ need, is known as Information Retrieval. This research uses Vector Space Model method in determining the similarity percentage of each student’s assignment. This research uses PHP programming and MySQL database. The finding is represented by ranking the similarity of document with query, with mean average precision value of 0,874. It shows how accurate the application with the examination done by the experts, which is gained from the evaluation with 5 queries that is compared to 25 samples of documents. If the number of counted assignments has higher similarity, thus the process of similarity counting needs more time, it depends on the assignment’s number which is submitted.


Sign in / Sign up

Export Citation Format

Share Document