scholarly journals Segmentation of Tomato Growing Truss a Depth Image Conversion Model Based on CycleGAN

Author(s):  
Dae-Hyun Jung ◽  
Cheoul Young Kim ◽  
Taek Sung Lee ◽  
Soo Hyun Park

Abstract Background: The truss on tomato plants is a group or cluster of smaller stems where flowers and fruit develop, while a growing truss is the most extended part of the stem. Because the state of the growing truss reacts sensitively to the surrounding environment, it is essential to control the growth in the early stages. With the recent development of IT and artificial intelligence technology in agriculture, a previous study developed a real-time acquisition and evaluation method for images using robots. Further, we used image processing to locate the growing truss and flowering rooms to extract growth information such as the height of the flower room and hard crab. Among the different vision algorithms, the CycleGAN algorithm was used to generate and transform unpaired images using generatively learning images. In this study, we developed a robot-based system for simultaneously acquiring RGB and depth images of the tomato growing truss and flower room groups.Results: The segmentation performance for approximately 35 samples was compared through the false negative (FN) and false positive (FP) indicators. For the depth camera image, we obtained FN as 17.55±3.01% and FP as 17.76±3.55%. Similarly, for CycleGAN, we obtained FN as approximately 19.24±1.45% and FP as 18.24±1.54%. As a result of image processing through depth image, IoU was 63.56 ± 8.44%, and when segmentation was performed through CycelGAN, IoU was 69.25 ± 4.42%, indicating that CycleGAN is advantageous in extracting the desired growing truss. Conclusions: The scannability was confirmed when the image scanning robot drove in a straight line through the plantation in the tomato greenhouse, which confirmed the on-site possibility of the image extraction technique using CycleGAN. In the future, the proposed approach is expected to be used in vision technology to scan the tomato growth indicators in greenhouses using an unmanned robot platform.

2017 ◽  
Vol 919 (1) ◽  
pp. 55-59
Author(s):  
O.V. Raskatkina

There is a method of using the corded plumb as vertical reference straight line, located in front of the objective of a digital photocamera in the article. When we take picture of the object under study, there will be this straight line in the photo, from which we can carry out all necessary measurements in the Paint system with the following conversion them into metric system. All possible variants of location of the reference straight line relative to it axis are considered by the example of the construction of the tower round shape and it is shown a method of heeling calculation by image processing results. Experimental research to determine the degree of influence of plumb location in the photo relative to it axe on the accuracy of the heeling determination was carried out by shooting the brick chimney with the 30 metres height when the plumb is located on the chimney axis and on different distance from the left and right of the axis. It is set in the result that the plumb location has influence on the accuracy of heeling determination. The optimal location is on the centre of the top section of the chimney and there is shown the method of accounting corrections due to inaccurate location.


Processes ◽  
2021 ◽  
Vol 9 (7) ◽  
pp. 1128
Author(s):  
Chern-Sheng Lin ◽  
Yu-Ching Pan ◽  
Yu-Xin Kuo ◽  
Ching-Kun Chen ◽  
Chuen-Lin Tien

In this study, the machine vision and artificial intelligence algorithms were used to rapidly check the degree of cooking of foods and avoid the over-cooking of foods. Using a smart induction cooker for heating, the image processing program automatically recognizes the color of the food before and after cooking. The new cooking parameters were used to identify the cooking conditions of the food when it is undercooked, cooked, and overcooked. In the research, the camera was used in combination with the software for development, and the real-time image processing technology was used to obtain the information of the color of the food, and through calculation parameters, the cooking status of the food was monitored. In the second year, using the color space conversion, a novel algorithm, and artificial intelligence, the foreground segmentation was used to separate the vegetables from the background, and the cooking ripeness, cooking unevenness, oil glossiness, and sauce absorption were calculated. The image color difference and the distribution were used to judge the cooking conditions of the food, so that the cooking system can identify whether or not to adopt partial tumbling, or to end a cooking operation. A novel artificial intelligence algorithm is used in the relative field, and the error rate can be reduced to 3%. This work will significantly help researchers working in the advanced cooking devices.


Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1299
Author(s):  
Honglin Yuan ◽  
Tim Hoogenkamp ◽  
Remco C. Veltkamp

Deep learning has achieved great success on robotic vision tasks. However, when compared with other vision-based tasks, it is difficult to collect a representative and sufficiently large training set for six-dimensional (6D) object pose estimation, due to the inherent difficulty of data collection. In this paper, we propose the RobotP dataset consisting of commonly used objects for benchmarking in 6D object pose estimation. To create the dataset, we apply a 3D reconstruction pipeline to produce high-quality depth images, ground truth poses, and 3D models for well-selected objects. Subsequently, based on the generated data, we produce object segmentation masks and two-dimensional (2D) bounding boxes automatically. To further enrich the data, we synthesize a large number of photo-realistic color-and-depth image pairs with ground truth 6D poses. Our dataset is freely distributed to research groups by the Shape Retrieval Challenge benchmark on 6D pose estimation. Based on our benchmark, different learning-based approaches are trained and tested by the unified dataset. The evaluation results indicate that there is considerable room for improvement in 6D object pose estimation, particularly for objects with dark colors, and photo-realistic images are helpful in increasing the performance of pose estimation algorithms.


2014 ◽  
Vol 1048 ◽  
pp. 173-177 ◽  
Author(s):  
Ying Mei Wang ◽  
Yan Mei Li ◽  
Wan Yue Hu

Fabric shape style is one of the most important conditions in textile appearance evaluation, and also the main factor influences customer purchasing psychology. At first, the previous fabric shape style evaluation methods are classified and summarized, measurement and evaluation method discussed from tactic and dynamic aspects. Then, companied with computer vision principle, a non-contact method for measuring fabric shape style was put forward. In this method, two high-speed CCD cameras were used to capture fabric movement dynamically, fabric sequences image were obtained in this process. Used the image processing technology include pretreatment and feature point matching to get 3D motion parameters, it can provide data supports for shape style evaluation.


Sensors ◽  
2021 ◽  
Vol 21 (4) ◽  
pp. 1356
Author(s):  
Linda Christin Büker ◽  
Finnja Zuber ◽  
Andreas Hein ◽  
Sebastian Fudickar

With approaches for the detection of joint positions in color images such as HRNet and OpenPose being available, consideration of corresponding approaches for depth images is limited even though depth images have several advantages over color images like robustness to light variation or color- and texture invariance. Correspondingly, we introduce High- Resolution Depth Net (HRDepthNet)—a machine learning driven approach to detect human joints (body, head, and upper and lower extremities) in purely depth images. HRDepthNet retrains the original HRNet for depth images. Therefore, a dataset is created holding depth (and RGB) images recorded with subjects conducting the timed up and go test—an established geriatric assessment. The images were manually annotated RGB images. The training and evaluation were conducted with this dataset. For accuracy evaluation, detection of body joints was evaluated via COCO’s evaluation metrics and indicated that the resulting depth image-based model achieved better results than the HRNet trained and applied on corresponding RGB images. An additional evaluation of the position errors showed a median deviation of 1.619 cm (x-axis), 2.342 cm (y-axis) and 2.4 cm (z-axis).


Mathematics ◽  
2021 ◽  
Vol 9 (21) ◽  
pp. 2815
Author(s):  
Shih-Hung Yang ◽  
Yao-Mao Cheng ◽  
Jyun-We Huang ◽  
Yon-Ping Chen

Automatic fingerspelling recognition tackles the communication barrier between deaf and hearing individuals. However, the accuracy of fingerspelling recognition is reduced by high intra-class variability and low inter-class variability. In the existing methods, regular convolutional kernels, which have limited receptive fields (RFs) and often cannot detect subtle discriminative details, are applied to learn features. In this study, we propose a receptive field-aware network with finger attention (RFaNet) that highlights the finger regions and builds inter-finger relations. To highlight the discriminative details of these fingers, RFaNet reweights the low-level features of the hand depth image with those of the non-forearm image and improves finger localization, even when the wrist is occluded. RFaNet captures neighboring and inter-region dependencies between fingers in high-level features. An atrous convolution procedure enlarges the RFs at multiple scales and a non-local operation computes the interactions between multi-scale feature maps, thereby facilitating the building of inter-finger relations. Thus, the representation of a sign is invariant to viewpoint changes, which are primarily responsible for intra-class variability. On an American Sign Language fingerspelling dataset, RFaNet achieved 1.77% higher classification accuracy than state-of-the-art methods. RFaNet achieved effective transfer learning when the number of labeled depth images was insufficient. The fingerspelling representation of a depth image can be effectively transferred from large- to small-scale datasets via highlighting the finger regions and building inter-finger relations, thereby reducing the requirement for expensive fingerspelling annotations.


2006 ◽  
Vol 20 (3) ◽  
pp. 265-273 ◽  
Author(s):  
Usman Ahmad ◽  
◽  
Fahmi Riadi ◽  
I Dewa Made Subrata ◽  
◽  
...  

The Lung Cancer is a most common cancer which causes of death to people. Early detection of this cancer will increase the survival rate. Usually, cancer detection is done manually by radiologists that had resulted in high rate of False Positive (FP) and False Negative (FN) test results. Currently Computed Tomography (CT) scan is used to scan the lung, which is much efficient than X-ray. In this proposed system a Computer Aided Detection (CADe) system for detecting lung cancer is used. This proposed system uses various image processing techniques to detect the lung cancer and also to classify the stages of lung cancer. Thus the rates of human errors are reduced in this system. As the result, the rate of obtaining False positive and (FP) False Negative (FN) has reduced. In this system, MATLAB have been used to process the image. Region growing algorithm is used to segment the ROI (Region of Interest). The SVM (Support Vector Machine) classifier is used to detect lung cancer and to identify the stages of lung cancer for the segmented ROI region. This proposed system produced 98.5 % accuracy when compared to other existing system


Author(s):  
Hyun Jun Park ◽  
Kwang Baek Kim

<p><span>Intel RealSense depth camera provides depth image using infrared projector and infrared camera. Using infrared radiation makes it possible to measure the depth with high accuracy, but the shadow of infrared radiation makes depth unmeasured regions. Intel RealSense SDK provides a postprocessing algorithm to correct it. However, this algorithm is not enough to be used and needs to be improved. Therefore, we propose a method to correct the depth image using image processing techniques. The proposed method corrects the depth using the adjacent depth information. Experimental results showed that the proposed method corrects the depth image more accurately than the Intel RealSense SDK.</span></p>


Sign in / Sign up

Export Citation Format

Share Document