Computer Vision for Object Recognition and Tracking Based on Raspberry Pi

2017 ◽  
pp. 3-14
Author(s):  
Ali A. Abed ◽  
Sara A. Rahman
2020 ◽  
Vol 67 (1) ◽  
pp. 133-141
Author(s):  
Dmitriy O. Khort ◽  
Aleksei I. Kutyrev ◽  
Igor G. Smirnov ◽  
Rostislav A. Filippov ◽  
Roman V. Vershinin

Technological capabilities of agricultural units cannot be optimally used without extensive automation of production processes and the use of advanced computer control systems. (Research purpose) To develop an algorithm for recognizing the coordinates of the location and ripeness of garden strawberries in different lighting conditions and describe the technological process of its harvesting in field conditions using a robotic actuator mounted on a self-propelled platform. (Materials and methods) The authors have developed a self-propelled platform with an automatic actuator for harvesting garden strawberry, which includes an actuator with six degrees of freedom, a co-axial gripper, mg966r servos, a PCA9685 controller, a Logitech HD C270 computer vision camera, a single-board Raspberry Pi 3 Model B+ computer, VL53L0X laser sensors, a SZBK07 300W voltage regulator, a Hubsan X4 Pro H109S Li-polymer battery. (Results and discussion) Using the Python programming language 3.7.2, the authors have developed a control algorithm for the automatic actuator, including operations to determine the X and Y coordinates of berries, their degree of maturity, as well as to calculate the distance to berries. It has been found that the effectiveness of detecting berries, their area and boundaries with a camera and the OpenCV library at the illumination of 300 Lux reaches 94.6 percent’s. With an increase in the robotic platform speed to 1.5 kilometre per hour and at the illumination of 300 Lux, the average area of the recognized berries decreased by 9 percent’s to 95.1 square centimeter, at the illumination of 200 Lux, the area of recognized berries decreased by 17.8 percent’s to 88 square centimeter, and at the illumination of 100 Lux, the area of recognized berries decreased by 36.4 percent’s to 76 square centimeter as compared to the real area of berries. (Conclusions) The authors have provided rationale for the technological process and developed an algorithm for harvesting garden strawberry using a robotic actuator mounted on a self-propelled platform. It has been proved that lighting conditions have a significant impact on the determination of the area, boundaries and ripeness of berries using a computer vision camera.


Fast track article for IS&T International Symposium on Electronic Imaging 2021: Intelligent Robotics and Industrial Applications using Computer Vision 2021 proceedings.


Author(s):  
Abd El Rahman Shabayek ◽  
Olivier Morel ◽  
David Fofi

For long time, it was thought that the sensing of polarization by animals is invariably related to their behavior, such as navigation and orientation. Recently, it was found that polarization can be part of a high-level visual perception, permitting a wide area of vision applications. Polarization vision can be used for most tasks of color vision including object recognition, contrast enhancement, camouflage breaking, and signal detection and discrimination. The polarization based visual behavior found in the animal kingdom is briefly covered. Then, the authors go in depth with the bio-inspired applications based on polarization in computer vision and robotics. The aim is to have a comprehensive survey highlighting the key principles of polarization based techniques and how they are biologically inspired.


2013 ◽  
pp. 896-926
Author(s):  
Mehrtash Harandi ◽  
Javid Taheri ◽  
Brian C. Lovell

Recognizing objects based on their appearance (visual recognition) is one of the most significant abilities of many living creatures. In this study, recent advances in the area of automated object recognition are reviewed; the authors specifically look into several learning frameworks to discuss how they can be utilized in solving object recognition paradigms. This includes reinforcement learning, a biologically-inspired machine learning technique to solve sequential decision problems and transductive learning, and a framework where the learner observes query data and potentially exploits its structure for classification. The authors also discuss local and global appearance models for object recognition, as well as how similarities between objects can be learnt and evaluated.


Author(s):  
Debi Prosad Dogra

Scene understanding and object recognition heavily depend on the success of visual attention guided salient region detection in images and videos. Therefore, summarizing computer vision techniques that take the help of visual attention models to accomplish video object recognition and tracking, can be helpful to the researchers of computer vision community. In this chapter, it is aimed to present a philosophical overview of the possible applications of visual attention models in the context of object recognition and tracking. At the beginning of this chapter, a brief introduction to various visual saliency models suitable for object recognition is presented, that is followed by discussions on possible applications of attention models on video object tracking. The chapter also provides a commentary on the existing techniques available on this domain and discusses some of their possible extensions. It is believed that, prospective readers will benefit since the chapter comprehensively guides a reader to understand the pros and cons of this particular topic.


Author(s):  
Yuexing Han ◽  
Bing Wang ◽  
Hideki Koike ◽  
Masanori Idesawa

One of the main goals of image understanding and computer vision applications is to recognize an object from various images. Object recognition has been deeply developed for the last three decades, and a lot of approaches have been proposed. Generally, these methods of object recognition can successfully achieve their goal by relying on a large quantity of data. However, if the observed objects are shown to diverse configurations, it is difficult to recognize them with a limited database. One has to prepare enough data to exactly recognize one object with multi-configurations, and it is hard work to collect enough data only for a single object. In this chapter, the authors will introduce an approach to recognize objects with multi-configurations using the shape space theory. Firstly, two sets of landmarks are obtained from two objects in two-dimensional images. Secondly, the landmarks represented as two points are projected into a pre-shape space. Then, a series of new intermediate data can be obtained from data models in the pre-shape space. Finally, object recognition can be achieved in the shape space with the shape space theory.


Sensors ◽  
2020 ◽  
Vol 20 (3) ◽  
pp. 613
Author(s):  
David Safadinho ◽  
João Ramos ◽  
Roberto Ribeiro ◽  
Vítor Filipe ◽  
João Barroso ◽  
...  

The capability of drones to perform autonomous missions has led retail companies to use them for deliveries, saving time and human resources. In these services, the delivery depends on the Global Positioning System (GPS) to define an approximate landing point. However, the landscape can interfere with the satellite signal (e.g., tall buildings), reducing the accuracy of this approach. Changes in the environment can also invalidate the security of a previously defined landing site (e.g., irregular terrain, swimming pool). Therefore, the main goal of this work is to improve the process of goods delivery using drones, focusing on the detection of the potential receiver. We developed a solution that has been improved along its iterative assessment composed of five test scenarios. The built prototype complements the GPS through Computer Vision (CV) algorithms, based on Convolutional Neural Networks (CNN), running in a Raspberry Pi 3 with a Pi NoIR Camera (i.e., No InfraRed—without infrared filter). The experiments were performed with the models Single Shot Detector (SSD) MobileNet-V2, and SSDLite-MobileNet-V2. The best results were obtained in the afternoon, with the SSDLite architecture, for distances and heights between 2.5–10 m, with recalls from 59%–76%. The results confirm that a low computing power and cost-effective system can perform aerial human detection, estimating the landing position without an additional visual marker.


Author(s):  
Kezhen Chen ◽  
Irina Rabkina ◽  
Matthew D. McLure ◽  
Kenneth D. Forbus

Deep learning systems can perform well on some image recognition tasks. However, they have serious limitations, including requiring far more training data than humans do and being fooled by adversarial examples. By contrast, analogical learning over relational representations tends to be far more data-efficient, requiring only human-like amounts of training data. This paper introduces an approach that combines automatically constructed qualitative visual representations with analogical learning to tackle a hard computer vision problem, object recognition from sketches. Results from the MNIST dataset and a novel dataset, the Coloring Book Objects dataset, are provided. Comparison to existing approaches indicates that analogical generalization can be used to identify sketched objects from these datasets with several orders of magnitude fewer examples than deep learning systems require.


Sign in / Sign up

Export Citation Format

Share Document