scholarly journals Advanced Fully Convolutional Networks for Agricultural Field Boundary Detection

2021 ◽  
Vol 13 (4) ◽  
pp. 722
Author(s):  
Alireza Taravat ◽  
Matthias P. Wagner ◽  
Rogerio Bonifacio ◽  
David Petit

Accurate spatial information of agricultural fields is important for providing actionable information to farmers, managers, and policymakers. On the other hand, the automated detection of field boundaries is a challenging task due to their small size, irregular shape and the use of mixed-cropping systems making field boundaries vaguely defined. In this paper, we propose a strategy for field boundary detection based on the fully convolutional network architecture called ResU-Net. The benefits of this model are two-fold: first, residual units ease training of deep networks. Second, rich skip connections within the network could facilitate information propagation, allowing us to design networks with fewer parameters but better performance in comparison with the traditional U-Net model. An extensive experimental analysis is performed over the whole of Denmark using Sentinel-2 images and comparing several U-Net and ResU-Net field boundary detection algorithms. The presented results show that the ResU-Net model has a better performance with an average F1 score of 0.90 and average Jaccard coefficient of 0.80 in comparison to the U-Net model with an average F1 score of 0.88 and an average Jaccard coefficient of 0.77.

2019 ◽  
Vol 12 (1) ◽  
pp. 59 ◽  
Author(s):  
Khairiya Mudrik Masoud ◽  
Claudio Persello ◽  
Valentyn A. Tolpekin

Boundaries of agricultural fields are important features necessary for defining the location, shape, and spatial extent of agricultural units. They are commonly used to summarize production statistics at the field level. In this study, we investigate the delineation of agricultural field boundaries (AFB) from Sentinel-2 satellite images acquired over the Flevoland province, the Netherlands, using a deep learning technique based on fully convolutional networks (FCNs). We designed a multiple dilation fully convolutional network (MD-FCN) for AFB detection from Sentinel-2 images at 10 m resolution. Furthermore, we developed a novel super-resolution semantic contour detection network (named SRC-Net) using a transposed convolutional layer in the FCN architecture to enhance the spatial resolution of the AFB output from 10 m to 5 m resolution. The SRC-Net also improves the AFB maps at 5 m resolution by exploiting the spatial-contextual information in the label space. The results of the proposed SRC-Net outperform alternative upsampling techniques and are only slightly inferior to the results of the MD-FCN for AFB detection from RapidEye images acquired at 5 m resolution.


2020 ◽  
Vol 12 (12) ◽  
pp. 1990
Author(s):  
Matthias P. Wagner ◽  
Natascha Oppelt

Field mapping and information on agricultural landscapes is of increasing importance for many applications. Monitoring schemes and national cadasters provide a rich source of information but their maintenance and regular updating is costly and labor-intensive. Automatized mapping of fields based on remote sensing imagery may aid in this task and allow for a faster and more regular observation. Although remote sensing has seen extensive use in agricultural research topics, such as plant health monitoring, crop type classification, yield prediction, and irrigation, field delineation and extraction has seen comparatively little research interest. In this study, we present a field boundary detection technique based on deep learning and a variety of image features, and combine it with the graph-based growing contours (GGC) method to extract agricultural fields in a study area in northern Germany. The boundary detection step only requires red, green, and blue (RGB) data and is therefore largely independent of the sensor used. We compare different image features based on color and luminosity information and evaluate their usefulness for the task of field boundary detection. A model based on texture metrics, gradient information, Hessian matrix eigenvalues, and local statistics showed good results with accuracies up to 88.2%, an area under the ROC curve (AUC) of up to 0.94, and F1 score of up to 0.88. The exclusive use of these universal image features may also facilitate transferability to other regions. We further present modifications to the GGC method intended to aid in upscaling of the method through process acceleration with a minimal effect on results. We combined the boundary detection results with the GGC method for field polygon extraction. Results were promising, with the new GGC version performing similarly or better than the original version while experiencing an acceleration of 1.3× to 2.3× on different subsets and input complexities. Further research may explore other applications of the GGC method outside agricultural remote sensing and field extraction.


Algorithms ◽  
2021 ◽  
Vol 14 (5) ◽  
pp. 144
Author(s):  
Yuexing Han ◽  
Xiaolong Li ◽  
Bing Wang ◽  
Lu Wang

Image segmentation plays an important role in the field of image processing, helping to understand images and recognize objects. However, most existing methods are often unable to effectively explore the spatial information in 3D image segmentation, and they neglect the information from the contours and boundaries of the observed objects. In addition, shape boundaries can help to locate the positions of the observed objects, but most of the existing loss functions neglect the information from the boundaries. To overcome these shortcomings, this paper presents a new cascaded 2.5D fully convolutional networks (FCNs) learning framework to segment 3D medical images. A new boundary loss that incorporates distance, area, and boundary information is also proposed for the cascaded FCNs to learning more boundary and contour features from the 3D medical images. Moreover, an effective post-processing method is developed to further improve the segmentation accuracy. We verified the proposed method on LITS and 3DIRCADb datasets that include the liver and tumors. The experimental results show that the performance of the proposed method is better than existing methods with a Dice Per Case score of 74.5% for tumor segmentation, indicating the effectiveness of the proposed method.


Author(s):  
Sophia Bano ◽  
Francisco Vasconcelos ◽  
Emmanuel Vander Poorten ◽  
Tom Vercauteren ◽  
Sebastien Ourselin ◽  
...  

Abstract Purpose Fetoscopic laser photocoagulation is a minimally invasive surgery for the treatment of twin-to-twin transfusion syndrome (TTTS). By using a lens/fibre-optic scope, inserted into the amniotic cavity, the abnormal placental vascular anastomoses are identified and ablated to regulate blood flow to both fetuses. Limited field-of-view, occlusions due to fetus presence and low visibility make it difficult to identify all vascular anastomoses. Automatic computer-assisted techniques may provide better understanding of the anatomical structure during surgery for risk-free laser photocoagulation and may facilitate in improving mosaics from fetoscopic videos. Methods We propose FetNet, a combined convolutional neural network (CNN) and long short-term memory (LSTM) recurrent neural network architecture for the spatio-temporal identification of fetoscopic events. We adapt an existing CNN architecture for spatial feature extraction and integrated it with the LSTM network for end-to-end spatio-temporal inference. We introduce differential learning rates during the model training to effectively utilising the pre-trained CNN weights. This may support computer-assisted interventions (CAI) during fetoscopic laser photocoagulation. Results We perform quantitative evaluation of our method using 7 in vivo fetoscopic videos captured from different human TTTS cases. The total duration of these videos was 5551 s (138,780 frames). To test the robustness of the proposed approach, we perform 7-fold cross-validation where each video is treated as a hold-out or test set and training is performed using the remaining videos. Conclusion FetNet achieved superior performance compared to the existing CNN-based methods and provided improved inference because of the spatio-temporal information modelling. Online testing of FetNet, using a Tesla V100-DGXS-32GB GPU, achieved a frame rate of 114 fps. These results show that our method could potentially provide a real-time solution for CAI and automating occlusion and photocoagulation identification during fetoscopic procedures.


Author(s):  
Qianmu Yuan ◽  
Jianwen Chen ◽  
Huiying Zhao ◽  
Yaoqi Zhou ◽  
Yuedong Yang

Abstract Motivation Protein–protein interactions (PPI) play crucial roles in many biological processes, and identifying PPI sites is an important step for mechanistic understanding of diseases and design of novel drugs. Since experimental approaches for PPI site identification are expensive and time-consuming, many computational methods have been developed as screening tools. However, these methods are mostly based on neighbored features in sequence, and thus limited to capture spatial information. Results We propose a deep graph-based framework deep Graph convolutional network for Protein–Protein-Interacting Site prediction (GraphPPIS) for PPI site prediction, where the PPI site prediction problem was converted into a graph node classification task and solved by deep learning using the initial residual and identity mapping techniques. We showed that a deeper architecture (up to eight layers) allows significant performance improvement over other sequence-based and structure-based methods by more than 12.5% and 10.5% on AUPRC and MCC, respectively. Further analyses indicated that the predicted interacting sites by GraphPPIS are more spatially clustered and closer to the native ones even when false-positive predictions are made. The results highlight the importance of capturing spatially neighboring residues for interacting site prediction. Availability and implementation The datasets, the pre-computed features, and the source codes along with the pre-trained models of GraphPPIS are available at https://github.com/biomed-AI/GraphPPIS. The GraphPPIS web server is freely available at https://biomed.nscc-gz.cn/apps/GraphPPIS. Supplementary information Supplementary data are available at Bioinformatics online.


2018 ◽  
Vol 10 (11) ◽  
pp. 1827 ◽  
Author(s):  
Ahram Song ◽  
Jaewan Choi ◽  
Youkyung Han ◽  
Yongil Kim

Hyperspectral change detection (CD) can be effectively performed using deep-learning networks. Although these approaches require qualified training samples, it is difficult to obtain ground-truth data in the real world. Preserving spatial information during training is difficult due to structural limitations. To solve such problems, our study proposed a novel CD method for hyperspectral images (HSIs), including sample generation and a deep-learning network, called the recurrent three-dimensional (3D) fully convolutional network (Re3FCN), which merged the advantages of a 3D fully convolutional network (FCN) and a convolutional long short-term memory (ConvLSTM). Principal component analysis (PCA) and the spectral correlation angle (SCA) were used to generate training samples with high probabilities of being changed or unchanged. The strategy assisted in training fewer samples of representative feature expression. The Re3FCN was mainly comprised of spectral–spatial and temporal modules. Particularly, a spectral–spatial module with a 3D convolutional layer extracts the spectral–spatial features from the HSIs simultaneously, whilst a temporal module with ConvLSTM records and analyzes the multi-temporal HSI change information. The study first proposed a simple and effective method to generate samples for network training. This method can be applied effectively to cases with no training samples. Re3FCN can perform end-to-end detection for binary and multiple changes. Moreover, Re3FCN can receive multi-temporal HSIs directly as input without learning the characteristics of multiple changes. Finally, the network could extract joint spectral–spatial–temporal features and it preserved the spatial structure during the learning process through the fully convolutional structure. This study was the first to use a 3D FCN and a ConvLSTM for the remote-sensing CD. To demonstrate the effectiveness of the proposed CD method, we performed binary and multi-class CD experiments. Results revealed that the Re3FCN outperformed the other conventional methods, such as change vector analysis, iteratively reweighted multivariate alteration detection, PCA-SCA, FCN, and the combination of 2D convolutional layers-fully connected LSTM.


2022 ◽  
Author(s):  
Ali Bahari Malayeri ◽  
Mohammad Bagher Khodabakhshi

Abstract Due to the importance of continuous monitoring of blood pressure (BP) in controlling hypertension, the topic of cuffless blood pressure (BP) estimation has been widely studied in recent years. A most important approach is to explore the nonlinear mapping between the recorded peripheral signals and the BP values which is usually conducted by deep neural networks. Because of the sequence-based pseudo periodic nature of peripheral signals such as photoplethysmogram (PPG), a proper estimation model needed to be equipped with the 1-dimensional (1-D) and recurrent layers. This, in turn, limits the usage of 2-dimensional (2-D) layers adopted in convolutional neural networks (CNN) for embedding spatial information in the model. In this study, considering the advantage of chaotic approaches, the recurrence characterization of peripheral signals was taken into account by a visual 2-D representation of PPG in phase space through fuzzy recurrence plot (FRP). FRP not only provides a beneficial framework for capturing the spatial properties of input signals but also creates a reliable approach for embedding the pseudo periodic properties to the neural models without using recurrent layers. Moreover, this study proposes a novel deep neural network architecture that combines the morphological features extracted simultaneously from two upgraded 1-D and 2-D CNNs capturing the temporal and spatial dependencies of PPGs in systolic and diastolic BP estimation. The model has been fed with the 1-D PPG sequences and the corresponding 2-D FRPs from two separate routes. The performance of the proposed framework was examined on the well-known public dataset, namely, Multi-Parameter Intelligent in Intensive Care II. Our scheme is analyzed and compared with the literature in terms of the requirements of the standards set by the British Hypertension Society (BHS) and the Association for the Advancement of Medical Instrumentation (AAMI). The proposed model met the AAMI requirements, and it achieved a grade of A as stated by the BHS standard. In addition, its mean absolute errors (MAE) and standard deviation for both systolic and diastolic blood pressure estimations were considerably low, 3.05±5.26 mmHg and 1.58±2.6 mmHg, in turn.


Geophysics ◽  
2021 ◽  
pp. 1-66
Author(s):  
Guanqun Sheng ◽  
Shuangyu Yang ◽  
Xiaolong Guo ◽  
Xingong Tang

Arrival-time picking of microseismic events is a critical procedure in microseismic data processing. However, as field monitoring data contain many microseismic events with low signal-to-noise ratios (SNRs), traditional arrival-time picking methods based on the instantaneous characteristics of seismic signals cannot meet the picking accuracy and efficiency requirements of microseismic monitoring owing to the large volume of monitoring data. Conversely, methods based on deep neural networks can significantly improve arrival-time picking accuracy and efficiency in low-SNR environments. Therefore, we propose a deep convolutional network that combines the U-net and DenseNet approaches to pick arrival times automatically. This novel network, called MSNet not only retains the spatial information of any input signal or profile based on the U-net, but also extracts and integrates more essential features of events and non-events through dense blocks, thereby further improving the picking accuracy and efficiency. An effective workflow is developed to verify the superiority of the proposed method. First, we describe the structure of MSNet and the workflow of the proposed picking method. Then, datasets are constructed using variable microseismic traces from field microseismic monitoring records and from the finite-difference forward modeling of microseismic data to train the network. Subsequently, hyperparameter tuning is conducted to optimize the MSNet. Finally, we test the MSNet using modeled signals with different SNRs and field microseismic data from different monitoring areas. By comparing the picking results of the proposed method with the results of U-net and short-term average and long-term average (STA/LTA) methods, the effectiveness of the proposed method is verified. The arrival picking results of synthetic data and microseismic field data show that the proposed network has increased adaptability and can achieve high accuracy for picking the arrival-time of microseismic events.


Author(s):  
Hang Gong ◽  
Shangdong Zheng ◽  
Zebin Wu ◽  
Yang Xu ◽  
Zhihui Wei ◽  
...  

The small defects in overhead catenary system (OCS) can result in long time delays, economic loss and even passenger injury. However, OCS images exhibit great variations with complex background and oblique views which pose a great challenge for small defects detection in high-speed rail system. In this paper, we propose the spatial-prior-guided attention for small object detection in OCS with two main advantages: (1) The spatial-prior is proposed to retain the spatial information between small defects and the electric components in OCS. (2) Based on spatial-prior, the spatial-prior-guided attention model (SAM) is designed to highlight useful information in the features and suppress redundant features response. SAM can model the spatial relations progressively and can be integrated with state-of-the-art feed-forward network architecture with end-to-end training fashion. We conduct extensive experiments on both Split pin datasets and PASCAL–VOC datasets and achieve 97.2% and 79.5% mAP values, respectively. All the experiments demonstrate the competitive performance of our method.


Sign in / Sign up

Export Citation Format

Share Document