scholarly journals TasselNetV2+: A Fast Implementation for High-Throughput Plant Counting From High-Resolution RGB Imagery

2020 ◽  
Vol 11 ◽  
Author(s):  
Hao Lu ◽  
Zhiguo Cao

Plant counting runs through almost every stage of agricultural production from seed breeding, germination, cultivation, fertilization, pollination to yield estimation, and harvesting. With the prevalence of digital cameras, graphics processing units and deep learning-based computer vision technology, plant counting has gradually shifted from traditional manual observation to vision-based automated solutions. One of popular solutions is a state-of-the-art object detection technique called Faster R-CNN where plant counts can be estimated from the number of bounding boxes detected. It has become a standard configuration for many plant counting systems in plant phenotyping. Faster R-CNN, however, is expensive in computation, particularly when dealing with high-resolution images. Unfortunately high-resolution imagery is frequently used in modern plant phenotyping platforms such as unmanned aerial vehicles, engendering inefficient image analysis. Such inefficiency largely limits the throughput of a phenotyping system. The goal of this work hence is to provide an effective and efficient tool for high-throughput plant counting from high-resolution RGB imagery. In contrast to conventional object detection, we encourage another promising paradigm termed object counting where plant counts are directly regressed from images, without detecting bounding boxes. In this work, by profiling the computational bottleneck, we implement a fast version of a state-of-the-art plant counting model TasselNetV2 with several minor yet effective modifications. We also provide insights why these modifications make sense. This fast version, TasselNetV2+, runs an order of magnitude faster than TasselNetV2, achieving around 30 fps on image resolution of 1980 × 1080, while it still retains the same level of counting accuracy. We validate its effectiveness on three plant counting tasks, including wheat ears counting, maize tassels counting, and sorghum heads counting. To encourage the use of this tool, our implementation has been made available online at https://tinyurl.com/TasselNetV2plus.

2021 ◽  
Author(s):  
Da-Ren Chen ◽  
Wei-Min Chiu

Abstract Machine learning techniques have been used to increase detection accuracy of cracks in road surfaces. Most studies failed to consider variable illumination conditions on the target of interest (ToI), and only focus on detecting the presence or absence of road cracks. This paper proposes a new road crack detection method, IlumiCrack, which integrates Gaussian mixture models (GMM) and object detection CNN models. This work provides the following contributions: 1) For the first time, a large-scale road crack image dataset with a range of illumination conditions (e.g., day and night) is prepared using a dashcam. 2) Based on GMM, experimental evaluations on 2 to 4 levels of brightness are conducted for optimal classification. 3) the IlumiCrack framework is used to integrate state-of-the-art object detecting methods with CNN to classify the road crack images into eight types with high accuracy. Experimental results show that IlumiCrack outperforms the state-of-the-art R-CNN object detection frameworks.


2018 ◽  
Vol 232 ◽  
pp. 04036
Author(s):  
Jun Yin ◽  
Huadong Pan ◽  
Hui Su ◽  
Zhonggeng Liu ◽  
Zhirong Peng

We propose an object detection method that predicts the orientation bounding boxes (OBB) to estimate objects locations, scales and orientations based on YOLO (You Only Look Once), which is one of the top detection algorithms performing well both in accuracy and speed. Horizontal bounding boxes(HBB), which are not robust to orientation variances, are used in the existing object detection methods to detect targets. The proposed orientation invariant YOLO (OIYOLO) detector can effectively deal with the bird’s eye viewpoint images where the orientation angles of the objects are arbitrary. In order to estimate the rotated angle of objects, we design a new angle loss function. Therefore, the training of OIYOLO forces the network to learn the annotated orientation angle of objects, making OIYOLO orientation invariances. The proposed approach that predicts OBB can be applied in other detection frameworks. In additional, to evaluate the proposed OIYOLO detector, we create an UAV-DAHUA datasets that annotated with objects locations, scales and orientation angles accurately. Extensive experiments conducted on UAV-DAHUA and DOTA datasets demonstrate that OIYOLO achieves state-of-the-art detection performance with high efficiency comparing with the baseline YOLO algorithms.


2020 ◽  
Vol 34 (07) ◽  
pp. 12959-12966
Author(s):  
Pengyu Zhao ◽  
Ansheng You ◽  
Yuanxing Zhang ◽  
Jiaying Liu ◽  
Kaigui Bian ◽  
...  

With the advance of omnidirectional panoramic technology, 360◦ imagery has become increasingly popular in the past few years. To better understand the 360◦ content, many works resort to the 360◦ object detection and various criteria have been proposed to bound the objects and compute the intersection-over-union (IoU) between bounding boxes based on the common equirectangular projection (ERP) or perspective projection (PSP). However, the existing 360◦ criteria are either inaccurate or inefficient for real-world scenarios. In this paper, we introduce a novel spherical criteria for fast and accurate 360◦ object detection, including both spherical bounding boxes and spherical IoU (SphIoU). Based on the spherical criteria, we propose a novel two-stage 360◦ detector, i.e., Reprojection R-CNN, by combining the advantages of both ERP and PSP, yielding efficient and accurate 360◦ object detection. To validate the design of spherical criteria and Reprojection R-CNN, we construct two unbiased synthetic datasets for training and evaluation. Experimental results reveal that compared with the existing criteria, the two-stage detector with spherical criteria achieves the best mAP results under the same inference speed, demonstrating that the spherical criteria can be more suitable for 360◦ object detection. Moreover, Reprojection R-CNN outperforms the previous state-of-the-art methods by over 30% on mAP with competitive speed, which confirms the efficiency and accuracy of the design.


2020 ◽  
Vol 34 (10) ◽  
pp. 13789-13790 ◽  
Author(s):  
Anurag Garg ◽  
Niket Tandon ◽  
Aparna S. Varde

Can we automatically predict failures of an object detection model on images from a target domain? We characterize errors of a state-of-the-art object detection model on the currently popular smart mobility domain, and find that a large number of errors can be identified using spatial commonsense. We propose øurmodel , a system that automatically identifies a large number of such errors based on commonsense knowledge. Our system does not require any new annotations and can still find object detection errors with high accuracy (more than 80% when measured by humans). This work lays the foundation to answer exciting research questions on domain adaptation including the ability to automatically create adversarial datasets for target domain.


Sensors ◽  
2021 ◽  
Vol 21 (10) ◽  
pp. 3374
Author(s):  
Hansen Liu ◽  
Kuangang Fan ◽  
Qinghua Ouyang ◽  
Na Li

To address the threat of drones intruding into high-security areas, the real-time detection of drones is urgently required to protect these areas. There are two main difficulties in real-time detection of drones. One of them is that the drones move quickly, which leads to requiring faster detectors. Another problem is that small drones are difficult to detect. In this paper, firstly, we achieve high detection accuracy by evaluating three state-of-the-art object detection methods: RetinaNet, FCOS, YOLOv3 and YOLOv4. Then, to address the first problem, we prune the convolutional channel and shortcut layer of YOLOv4 to develop thinner and shallower models. Furthermore, to improve the accuracy of small drone detection, we implement a special augmentation for small object detection by copying and pasting small drones. Experimental results verify that compared to YOLOv4, our pruned-YOLOv4 model, with 0.8 channel prune rate and 24 layers prune, achieves 90.5% mAP and its processing speed is increased by 60.4%. Additionally, after small object augmentation, the precision and recall of the pruned-YOLOv4 almost increases by 22.8% and 12.7%, respectively. Experiment results verify that our pruned-YOLOv4 is an effective and accurate approach for drone detection.


2019 ◽  
Vol 77 (4) ◽  
pp. 1427-1439 ◽  
Author(s):  
Qiong Li ◽  
Xin Sun ◽  
Junyu Dong ◽  
Shuqun Song ◽  
Tongtong Zhang ◽  
...  

Abstract Phytoplankton plays an important role in marine ecological environment and aquaculture. However, the recognition and detection of phytoplankton rely on manual operations. As the foundation of achieving intelligence and releasing human labour, a phytoplankton microscopic image dataset PMID2019 for phytoplankton automated detection is presented. The PMID2019 dataset contains 10 819 phytoplankton microscopic images of 24 different categories. We leverage microscopes to collect images of phytoplankton in the laboratory environment. Each object in the images is manually labelled with a bounding box and category of ground-truth. In addition, living cells move quickly making it difficult to capture images of them. In order to generalize the dataset for in situ applications, we further utilize Cycle-GAN to achieve the domain migration between dead and living cell samples. We built a synthetic dataset to generate the corresponding living cell samples from the original dead ones. The PMID2019 dataset will not only benefit the development of phytoplankton microscopic vision technology in the future, but also can be widely used to assess the performance of the state-of-the-art object detection algorithms for phytoplankton recognition. Finally, we illustrate the performances of some state-of-the-art object detection algorithms, which may provide new ideas for monitoring marine ecosystems.


Informatics ◽  
2020 ◽  
Vol 17 (2) ◽  
pp. 7-16
Author(s):  
R. P. Bohush ◽  
I. Yu. Zakharava ◽  
S. V. Ablameyko

In the paper the algorithm for object detection in high resolution images is proposed. The approach uses multiscale image representation followed by block processing with the overlapping value. For each block the object detection with convolutional neural network was performed. Number of pyramid layers is limited by the Convolutional Neural Network layer size and input image resolution. Overlapping blocks splitting to improve the classification and detection accuracy is performed on each layer of pyramid except the highest one. Detected areas are merged into one if they have high overlapping value and the same class. Experimental results for the algorithm are presented in the paper.


2020 ◽  
Vol 8 (1) ◽  
pp. 53-62
Author(s):  
Alexander Leipnitz ◽  
Tilo Strutz ◽  
Oliver Jokisch

The automated operation of robots and flying drones is coupled to high security requirements with respect to humans and environment. Sometimes, persons have to be detected from a long distance or high altitude to allow the autonomous system an adequate and timely response. State-of-the-art Convolutional Neural Networks (CNNs) enable high object detection rates for different image data but only within their respective training, validation and test datasets. Recent studies show the limited generalization ability of CNNs for unknown data, even with merely small image changes. A typical source of such problems is the varying resolution of input images and the inevitable scaling of them to match the input-layer size of the network model. While modern cameras are able to capture high-resolution images of humans also from a longer distance, the practical input-layer sizes of networks are comparably small. Hence, we investigate the reliability of a network architecture for human detection with respect to such input-scaling effects. The popular VisDrone dataset with its varying image resolution and many relatively small depictions of humans is surveyed as well as the high-resolution AgriDrone image data from an agricultural context. Our results show that the object detection rate depends on the image scaling factor as well as on the relative size of persons. An enlarged input-layer size of the network can only partially contribute to counteract the observed effects. In addition, the detection algorithm becomes computationally more expensive by the increased effort.


Sign in / Sign up

Export Citation Format

Share Document