scholarly journals Computer Vision System for Expressing Texture Using Sound-Symbolic Words

2021 ◽  
Vol 12 ◽  
Author(s):  
Koichi Yamagata ◽  
Jinhwan Kwon ◽  
Takuya Kawashima ◽  
Wataru Shimoda ◽  
Maki Sakamoto

The major goals of texture research in computer vision are to understand, model, and process texture and ultimately simulate human visual information processing using computer technologies. The field of computer vision has witnessed remarkable advancements in material recognition using deep convolutional neural networks (DCNNs), which have enabled various computer vision applications, such as self-driving cars, facial and gesture recognition, and automatic number plate recognition. However, for computer vision to “express” texture like human beings is still difficult because texture description has no correct or incorrect answer and is ambiguous. In this paper, we develop a computer vision method using DCNN that expresses texture of materials. To achieve this goal, we focus on Japanese “sound-symbolic” words, which can describe differences in texture sensation at a fine resolution and are known to have strong and systematic sensory-sound associations. Because the phonemes of Japanese sound-symbolic words characterize categories of texture sensations, we develop a computer vision method to generate the phonemes and structure comprising sound-symbolic words that probabilistically correspond to the input images. It was confirmed that the sound-symbolic words output by our system had about 80% accuracy rate in our evaluation.

2021 ◽  
Vol 2135 (1) ◽  
pp. 012002
Author(s):  
Holman Montiel ◽  
Fernando Martínez ◽  
Fredy Martínez

Abstract Autonomous mobility remains an open research problem in robotics. This is a complex problem that has its characteristics according to the type of task and environment intended for the robot’s activity. Service robotics has in this sense problems that have not been solved satisfactorily. These robots must interact with human beings in environments designed for human beings, which implies that one of the basic sensors for structuring motion control and navigation schemes are those that replicate the human optical sense. In their normal activity, robots are expected to interpret visual information in the environment while following a certain motion policy that allows them to move from one point to another in the environment, consistent with their tasks. A good optical sensing system can be structured around digital cameras, with which it can apply visual identification routines of both the trajectory and its environment. This research proposes a parallel control scheme (with two loops) for the definition of movements of a service robot from images. On the one hand, there is a control loop based on a visual memory strategy using a convolutional neural network. This system contemplates a deep learning model that is trained from images of the environment containing characteristic elements of the navigation environment (various types of obstacles and different cases of free trajectories with and without navigation path). To this first loop is connected in parallel a second loop in charge of defining the specific distances to the obstacles using a stereo vision system. The objective of this parallel loop is to quickly identify the obstacle points in front of the robot from the images using a bacterial interaction model. These two loops form an information feedback motion control framework that quickly analyzes the environment and defines motion strategies from digital images, achieving real-time control driven by visual information. Among the advantages of our scheme are the low processing and memory costs in the robot, and the no need to modify the environment to facilitate the navigation of the robot. The performance of the system is validated by simulation and laboratory experiments.


2020 ◽  
Vol 40 (1) ◽  
pp. 21
Author(s):  
Ferlando Jubelito Simanungkalit ◽  
Rosnawyta Simanjuntak

Color had a correlation with physical appearance, nutritional and chemical content as well as sensory properties which determine the quality of agricultural products and foods. Conventional color measurements were performed destructively using laboratory equipment. Therefore, color measurement methods of agricultural products were needed more quickly, accurately and non-destructively. This study aimed to develop a Computer Vision System (CVS) that can be used as a tool to measure the color of fruits. The designed CVS consists of a 60x60x60 cm black mini photo studio; a pair 15 watt LED lighting, sony α6000 digital camera, a set of laptop and an image processing software applications. Image processing software was programmed using VB.Net 2008 programming language. The developed CVS was calibrated using 24 color charts Macbeth Colorchecker (Gretag-Macbeth, USA). The calibration results of 24 color chart of Macbeth Colorchecker was resulted in a MAPE (Mean Absolute Percentage Error) value of component R / Red = 0%; G / Green = 0% and B / Blue = 0,5%; with 99% accuracy rate. In color measurement, the developed CVS had a 95% accuracy rate.


Author(s):  
Bibhu Prasad ◽  
Ashima Sindhu Mohanty ◽  
Ami Kumar Parida

We synthetically applied computer vision, genetic algorithm and artificial neural network technology to automatically identify the vegetables (tomatoes) that had physiological diseases. Initially tomatoes’ images were captured through a computer vision system. Then to identify cavernous tomatoes, we analyzed the roundness and detected deformed tomatoes by applying the variation of vegetable’s diameter. Later, we used a Genetic Algorithm (GA) based artificial neural network (ANN). Experiments show that the above methods can accurately identify vegetables’ shapes and meet requests of classification; the accuracy rate for the identification for vegetables with physiological diseases was up to 100%. [Nature and Science. 2005; 3(2):52-58].


Author(s):  
Konstantin Dergachov ◽  
Anatolii Kulik ◽  
Anatolii Zymovin

In this chapter, the authors present an approach to the extrinsic environs diagnostics based on using visual information collected by autonomous robots. The possibility of utilizing a computer vision for the purpose of rational control implementation in the condition of the full or partial uncertainty is investigated. In the study, the combined hardware and software computer vision tools were verified. The models, algorithms, and codes for solving the local tasks of obstacle identification and mutual location kinematic parameters estimation have been developed. A series of computational and in-kind experiments that illustrate a practical possibility of implementing the navigational environment diagnosis is carried out with the aim to select a rational flight path.


2018 ◽  
Vol 1 (2) ◽  
pp. 17-23
Author(s):  
Takialddin Al Smadi

This survey outlines the use of computer vision in Image and video processing in multidisciplinary applications; either in academia or industry, which are active in this field.The scope of this paper covers the theoretical and practical aspects in image and video processing in addition of computer vision, from essential research to evolution of application.In this paper a various subjects of image processing and computer vision will be demonstrated ,these subjects are spanned from the evolution of mobile augmented reality (MAR) applications, to augmented reality under 3D modeling and real time depth imaging, video processing algorithms will be discussed to get higher depth video compression, beside that in the field of mobile platform an automatic computer vision system for citrus fruit has been implemented ,where the Bayesian classification with Boundary Growing to detect the text in the video scene. Also the paper illustrates the usability of the handed interactive method to the portable projector based on augmented reality.   © 2018 JASET, International Scholars and Researchers Association


2021 ◽  
Vol 8 (1) ◽  
Author(s):  
Eslam Mounier ◽  
Bassem Abdullah ◽  
Hani Mahdi ◽  
Seif Eldawlatly

AbstractThe Lateral Geniculate Nucleus (LGN) represents one of the major processing sites along the visual pathway. Despite its crucial role in processing visual information and its utility as one target for recently developed visual prostheses, it is much less studied compared to the retina and the visual cortex. In this paper, we introduce a deep learning encoder to predict LGN neuronal firing in response to different visual stimulation patterns. The encoder comprises a deep Convolutional Neural Network (CNN) that incorporates visual stimulus spatiotemporal representation in addition to LGN neuronal firing history to predict the response of LGN neurons. Extracellular activity was recorded in vivo using multi-electrode arrays from single units in the LGN in 12 anesthetized rats with a total neuronal population of 150 units. Neural activity was recorded in response to single-pixel, checkerboard and geometrical shapes visual stimulation patterns. Extracted firing rates and the corresponding stimulation patterns were used to train the model. The performance of the model was assessed using different testing data sets and different firing rate windows. An overall mean correlation coefficient between the actual and the predicted firing rates of 0.57 and 0.7 was achieved for the 10 ms and the 50 ms firing rate windows, respectively. Results demonstrate that the model is robust to variability in the spatiotemporal properties of the recorded neurons outperforming other examined models including the state-of-the-art Generalized Linear Model (GLM). The results indicate the potential of deep convolutional neural networks as viable models of LGN firing.


Minerals ◽  
2021 ◽  
Vol 11 (8) ◽  
pp. 791
Author(s):  
Sufei Zhang ◽  
Ying Guo

This paper introduces computer vision systems (CVSs), which provides a new method to measure gem colour, and compares CVS and colourimeter (CM) measurements of jadeite-jade colour in the CIELAB space. The feasibility of using CVS for jadeite-jade colour measurement was verified by an expert group test and a reasonable regression model in an experiment involving 111 samples covering almost all jadeite-jade colours. In the expert group test, more than 93.33% of CVS images are considered to have high similarities with real objects. Comparing L*, a*, b*, C*, h, and ∆E* (greater than 10) from CVS and CM tests indicate that significant visual differences exist between the measured colours. For a*, b*, and h, the R2 of the regression model for CVS and CM was 90.2% or more. CVS readings can be used to predict the colour value measured by CM, which means that CVS technology can become a practical tool to detect the colour of jadeite-jade.


2021 ◽  
pp. 105084
Author(s):  
Bojana Milovanovic ◽  
Ilija Djekic ◽  
Jelena Miocinovic ◽  
Bartosz G. Solowiej ◽  
Jose M. Lorenzo ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document