scholarly journals Drone Image Segmentation Using Machine and Deep Learning for Mapping Raised Bog Vegetation Communities

2020 ◽  
Vol 12 (16) ◽  
pp. 2602 ◽  
Author(s):  
Saheba Bhatnagar ◽  
Laurence Gill ◽  
Bidisha Ghosh

The application of drones has recently revolutionised the mapping of wetlands due to their high spatial resolution and the flexibility in capturing images. In this study, the drone imagery was used to map key vegetation communities in an Irish wetland, Clara Bog, for the spring season. The mapping, carried out through image segmentation or semantic segmentation, was performed using machine learning (ML) and deep learning (DL) algorithms. With the aim of identifying the most appropriate, cost-efficient, and accurate segmentation method, multiple ML classifiers and DL models were compared. Random forest (RF) was identified as the best pixel-based ML classifier, which provided good accuracy (≈85%) when used in conjunction graph cut algorithm for image segmentation. Amongst the DL networks, a convolutional neural network (CNN) architecture in a transfer learning framework was utilised. A combination of ResNet50 and SegNet architecture gave the best semantic segmentation results (≈90%). The high accuracy of DL networks was accompanied with significantly larger labelled training dataset, computation time and hardware requirements compared to ML classifiers with slightly lower accuracy. For specific applications such as wetland mapping where networks are required to be trained for each different site, topography, season, and other atmospheric conditions, ML classifiers proved to be a more pragmatic choice.

Sensors ◽  
2021 ◽  
Vol 21 (11) ◽  
pp. 3813
Author(s):  
Athanasios Anagnostis ◽  
Aristotelis C. Tagarakis ◽  
Dimitrios Kateris ◽  
Vasileios Moysiadis ◽  
Claus Grøn Sørensen ◽  
...  

This study aimed to propose an approach for orchard trees segmentation using aerial images based on a deep learning convolutional neural network variant, namely the U-net network. The purpose was the automated detection and localization of the canopy of orchard trees under various conditions (i.e., different seasons, different tree ages, different levels of weed coverage). The implemented dataset was composed of images from three different walnut orchards. The achieved variability of the dataset resulted in obtaining images that fell under seven different use cases. The best-trained model achieved 91%, 90%, and 87% accuracy for training, validation, and testing, respectively. The trained model was also tested on never-before-seen orthomosaic images or orchards based on two methods (oversampling and undersampling) in order to tackle issues with out-of-the-field boundary transparent pixels from the image. Even though the training dataset did not contain orthomosaic images, it achieved performance levels that reached up to 99%, demonstrating the robustness of the proposed approach.


2021 ◽  
Author(s):  
Neeraj Kumar Rathore ◽  
Varshali Jaiswal ◽  
Varsha Sharma ◽  
Sunita Varma

Abstract Deep-Convolution Neural Network (CNN) is the branch of computer science. Deep Learning CNN is a methodology that teaches computer systems to do what comes naturally to humans. It is a method that learns by example and experience. This is a heuristic-based method to solve computationally exhaustive problems that are not resolved in a polynomial computation time like NP-Hard problems. The purpose of this research is to develop a hybrid methodology for the detection and segmentation of flower images that utilize the extension of the deep CNN. The plant, leaf, and flower image detection are the most challenging issues due to a wide variety of classes, based on their amount of texture, color distinctiveness, shape distinctiveness, and different size. The proposed methodology is implemented in Matlab with deep learning Tool Box and the dataset of flower image is taken from Kaggle with five different classes like daisy, dandelion, rose, tulip, and sunflower. This methodology takes an input of different flower images from data sets, then converts it from RGB (Red, Green, Blue) color model to the L*a*b color model. L*a*b has reduced the effort of image segmentation. The flower image segmentation has been performed by the canny edge detection algorithm which provided better results. The implemented extended deep learning convolution neural network can accurately recognize varieties of flower images. The learning accuracy of the proposed hybrid method is up to 98% that is maximizing up to + 1.89% from state of the art.


2021 ◽  
Vol 10 (8) ◽  
pp. 551
Author(s):  
Jiaxin Zhang ◽  
Tomohiro Fukuda ◽  
Nobuyoshi Yabuki

Precise measuring of urban façade color is necessary for urban color planning. The existing manual methods of measuring building façade color are limited by time and labor costs and hardly carried out on a city scale. These methods also make it challenging to identify the role of the building function in controlling and guiding urban color planning. This paper explores a city-scale approach to façade color measurement with building functional classification using state-of-the-art deep learning techniques and street view images. Firstly, we used semantic segmentation to extract building façades and conducted the color calibration of the photos for pre-processing the collected street view images. Then, we proposed a color chart-based façade color measurement method and a multi-label deep learning-based building classification method. Next, the field survey data were used as the ground truth to verify the accuracy of the façade color measurement and building function classification. Finally, we applied our approach to generate façade color distribution maps with the building classification for three metropolises in China, and the results proved the transferability and effectiveness of the scheme. The proposed approach can provide city managers with an overall perception of urban façade color and building function across city-scale areas in a cost-efficient way, contributing to data-driven decision making for urban analytics and planning.


2021 ◽  
Vol 21 (1) ◽  
Author(s):  
Dominik Müller ◽  
Frank Kramer

Abstract Background The increased availability and usage of modern medical imaging induced a strong need for automatic medical image segmentation. Still, current image segmentation platforms do not provide the required functionalities for plain setup of medical image segmentation pipelines. Already implemented pipelines are commonly standalone software, optimized on a specific public data set. Therefore, this paper introduces the open-source Python library MIScnn. Implementation The aim of MIScnn is to provide an intuitive API allowing fast building of medical image segmentation pipelines including data I/O, preprocessing, data augmentation, patch-wise analysis, metrics, a library with state-of-the-art deep learning models and model utilization like training, prediction, as well as fully automatic evaluation (e.g. cross-validation). Similarly, high configurability and multiple open interfaces allow full pipeline customization. Results Running a cross-validation with MIScnn on the Kidney Tumor Segmentation Challenge 2019 data set (multi-class semantic segmentation with 300 CT scans) resulted into a powerful predictor based on the standard 3D U-Net model. Conclusions With this experiment, we could show that the MIScnn framework enables researchers to rapidly set up a complete medical image segmentation pipeline by using just a few lines of code. The source code for MIScnn is available in the Git repository: https://github.com/frankkramer-lab/MIScnn.


Sensors ◽  
2021 ◽  
Vol 21 (6) ◽  
pp. 2077 ◽  
Author(s):  
Shih-Yu Chen ◽  
Chinsu Lin ◽  
Guan-Jie Li ◽  
Yu-Chun Hsu ◽  
Keng-Hao Liu

The life cycle of leaves, from sprout to senescence, is the phenomenon of regular changes such as budding, branching, leaf spreading, flowering, fruiting, leaf fall, and dormancy due to seasonal climate changes. It is the effect of temperature and moisture in the life cycle on physiological changes, so the detection of newly grown leaves (NGL) is helpful for the estimation of tree growth and even climate change. This study focused on the detection of NGL based on deep learning convolutional neural network (CNN) models with sparse enhancement (SE). As the NGL areas found in forest images have similar sparse characteristics, we used a sparse image to enhance the signal of the NGL. The difference between the NGL and the background could be further improved. We then proposed hybrid CNN models that combined U-net and SegNet features to perform image segmentation. As the NGL in the image were relatively small and tiny targets, in terms of data characteristics, they also belonged to the problem of imbalanced data. Therefore, this paper further proposed 3-Layer SegNet, 3-Layer U-SegNet, 2-Layer U-SegNet, and 2-Layer Conv-U-SegNet architectures to reduce the pooling degree of traditional semantic segmentation models, and used a loss function to increase the weight of the NGL. According to the experimental results, our proposed algorithms were indeed helpful for the image segmentation of NGL and could achieve better kappa results by 0.743.


Author(s):  
Ganesh R. Padalkar ◽  
Madhuri B. Khambete

Semantic segmentation is a pre-processing step in computer vision-based applications. It is the task of assigning a predefined class label to every pixel of an image. Several supervised and unsupervised algorithms are available to classify pixels of an image into predefined object classes. The algorithms, such as random forest and SVM are used to obtain the semantic segmentation. Recently, convolutional neural network (CNN)-based architectures have become popular for the tasks of object detection, object recognition, and segmentation. These deep architectures perform semantic segmentation with far better accuracy than the algorithms that were used earlier. CNN-based deep learning architectures require a large dataset for training. In real life, some of the applications may not have sufficient good quality samples for training of deep learning architectures e.g. medical applications. Such a requirement initiated a need to have a technique of effective training of deep learning architecture in case of a very small dataset. Class imbalance is another challenge in the process of training deep learning architecture. Due to class imbalance, the classifier overclassifies classes with large samples. In this paper, the challenge of training a deep learning architecture with a small dataset and class imbalance is addressed by novel fusion-based semantic segmentation technique which improves segmentation of minor and major classes.


2021 ◽  
Vol 13 (13) ◽  
pp. 2578
Author(s):  
Samir Touzani ◽  
Jessica Granderson

Advances in machine learning and computer vision, combined with increased access to unstructured data (e.g., images and text), have created an opportunity for automated extraction of building characteristics, cost-effectively, and at scale. These characteristics are relevant to a variety of urban and energy applications, yet are time consuming and costly to acquire with today’s manual methods. Several recent research studies have shown that in comparison to more traditional methods that are based on features engineering approach, an end-to-end learning approach based on deep learning algorithms significantly improved the accuracy of automatic building footprint extraction from remote sensing images. However, these studies used limited benchmark datasets that have been carefully curated and labeled. How the accuracy of these deep learning-based approach holds when using less curated training data has not received enough attention. The aim of this work is to leverage the openly available data to automatically generate a larger training dataset with more variability in term of regions and type of cities, which can be used to build more accurate deep learning models. In contrast to most benchmark datasets, the gathered data have not been manually curated. Thus, the training dataset is not perfectly clean in terms of remote sensing images exactly matching the ground truth building’s foot-print. A workflow that includes data pre-processing, deep learning semantic segmentation modeling, and results post-processing is introduced and applied to a dataset that include remote sensing images from 15 cities and five counties from various region of the USA, which include 8,607,677 buildings. The accuracy of the proposed approach was measured on an out of sample testing dataset corresponding to 364,000 buildings from three USA cities. The results favorably compared to those obtained from Microsoft’s recently released US building footprint dataset.


2021 ◽  
Vol 11 (19) ◽  
pp. 9180
Author(s):  
Siangruei Wu ◽  
Yihong Wu ◽  
Haoyun Chang ◽  
Florence T. Su ◽  
Hengchun Liao ◽  
...  

Semantic segmentation of medical images with deep learning models is rapidly being developed. In this study, we benchmarked state-of-the-art deep learning segmentation algorithms on our clinical stereotactic radiosurgery dataset. The dataset consists of 1688 patients with various brain lesions (pituitary tumors, meningioma, schwannoma, brain metastases, arteriovenous malformation, and trigeminal neuralgia), and we divided the dataset into a training set (1557 patients) and test set (131 patients). This study demonstrates the strengths and weaknesses of deep-learning algorithms in a fairly practical scenario. We compared the model performances concerning their sampling method, model architecture, and the choice of loss functions, identifying suitable settings for their applications and shedding light on the possible improvements. Evidence from this study led us to conclude that deep learning could be promising in assisting the segmentation of brain lesions even if the training dataset was of high heterogeneity in lesion types and sizes.


2021 ◽  
Author(s):  
SHOGO ARAI ◽  
ZHUANG FENG ◽  
Fuyuki Tokuda ◽  
Adam Purnomo ◽  
Kazuhiro Kosuge

<div>This paper proposes a deep learning-based fast grasp detection method with a small dataset for robotic bin-picking. We consider the problem of grasping stacked up mechanical parts on a planar workspace using a parallel gripper. In this paper, we use a deep neural network to solve the problem with a single depth image. To reduce the computation time, we propose an edge-based algorithm to generate potential grasps. Then, a convolutional neural network (CNN) is applied to evaluate the robustness of all potential grasps for bin-picking. Finally, the proposed method ranks them and the object is grasped by using the grasp with the highest score. In bin-picking experiments, we evaluate the proposed method with a 7-DOF manipulator using textureless mechanical parts with complex shapes. The success ratio of grasping is 97%, and the average computation time of CNN inference is less than 0.23[s] on a laptop PC without a GPU array. In addition, we also confirm that the proposed method can be applied to unseen objects which are not included in the training dataset. </div>


Sign in / Sign up

Export Citation Format

Share Document