Gaze-Contingent Visual Presentation Based on Remote Saccade Detection

2007 ◽  
Vol 16 (2) ◽  
pp. 224-234 ◽  
Author(s):  
Junji Watanabe ◽  
Hideyuki Ando ◽  
Taro Maeda ◽  
Susumu Tachi

Pursuing new display techniques based on insights into human visual perception can reveal new possibilities for visual information devices. Here, we propose a novel information presentation technique that exploits the perceptional features during rapid eye movements called saccades by using a fast remote eye-measuring method. When light sources are fixed on a vertical line, and the flashing pattern is changed quickly during a horizontal saccade, 2D images can be perceived due to spatio-temporal integration in the human vision system. We use this phenomenon to present 2D images with only one-dimensional light sources, and to show these images even in midair. The flashing cycle and flash timing of light sources are important elements in developing the design theory for this display technique. The flashing cycle determines the maximum resolution of a perceived 2D image. The flash timing is a crucial issue for our purpose because 2D images are perceived only when the timing of the saccade coincides with the flash timing. Therefore, in this paper, we describe the relationship between a flashing cycle and the maximum resolution of a perceived 2D image, and then propose a concise saccade detection method. By using this method, saccades can be detected and the light sources can be flashed within the saccade interval as it occurs in real time, and 2D images can be successfully presented.

Author(s):  
Basavaraj S Anami ◽  
Dayanand G Savakar

The paper presents an artificial neural network based approach to corroborate the effect of variations in illumination, distance of image acquisition and size on recognition and classification accuracies of bulk food grain image samples. Different food grains samples like Wheat, Groundnut, Green gram and Jowar are considered. The image samples are taken by varying acquisition distances, illumination and sizes. The natural light source and the minimum distance of 40 centimeters are ideal for image acquisition compared to other light sources and varying distance of image acquisition. The experimental results have shown a reduction in accuracy in all the cases other than natural lighting condition and fixed distance of acquisition. The same is true with human vision system too. Variations in image sizes also affect the recognition and classification accuracies. Hence, it is inferred that any variation in the size of the image and the distance of acquisition has an impact on the accuracy of recognition of food grain samples.


Sensors ◽  
2021 ◽  
Vol 21 (2) ◽  
pp. 343
Author(s):  
Kim Bjerge ◽  
Jakob Bonde Nielsen ◽  
Martin Videbæk Sepstrup ◽  
Flemming Helsing-Nielsen ◽  
Toke Thomas Høye

Insect monitoring methods are typically very time-consuming and involve substantial investment in species identification following manual trapping in the field. Insect traps are often only serviced weekly, resulting in low temporal resolution of the monitoring data, which hampers the ecological interpretation. This paper presents a portable computer vision system capable of attracting and detecting live insects. More specifically, the paper proposes detection and classification of species by recording images of live individuals attracted to a light trap. An Automated Moth Trap (AMT) with multiple light sources and a camera was designed to attract and monitor live insects during twilight and night hours. A computer vision algorithm referred to as Moth Classification and Counting (MCC), based on deep learning analysis of the captured images, tracked and counted the number of insects and identified moth species. Observations over 48 nights resulted in the capture of more than 250,000 images with an average of 5675 images per night. A customized convolutional neural network was trained on 2000 labeled images of live moths represented by eight different classes, achieving a high validation F1-score of 0.93. The algorithm measured an average classification and tracking F1-score of 0.71 and a tracking detection rate of 0.79. Overall, the proposed computer vision system and algorithm showed promising results as a low-cost solution for non-destructive and automatic monitoring of moths.


2012 ◽  
Vol 726 ◽  
pp. 226-232 ◽  
Author(s):  
Tomasz Giesko

The article presents a dual-camera vision system for fatigue monitoring composed of a vision unit, a camera positioning set and a computer unit. Vision modules are mounted onto the 4DOF positioning sets, which allows for an easy determination of the position of the camera in relation to the sample. The application of motorized measurement lenses with changeable configuration, thanks to the alteration of the distance of observation and the vision angle, enables the adaptation of the system to different scales of observation of the fatigue processes in the specimen surface. Automatic focus setting is realised with the use of the implemented algorithm. The software developed allows for the analysis of fatigue fracture for two 2D images or the 3D stereovision image.


Author(s):  
Д.А. Смирнов ◽  
В.Г. Бондарев ◽  
А.В. Николенко

Проведен краткий анализ как отечественных, так и зарубежных систем межсамолетной навигации. В ходе анализа были отражены недостатки систем межсамолетной навигации и представлен актуальный подход повышения точности системы навигации за счет применения системы технического зрения. Для определения местоположения ведущего самолета предлагается рассмотреть в качестве измерительного комплекса систему технического зрения, которая способна решать большой круг задач на различных этапах, в частности, и полет строем. Систему технического зрения предлагается установить на ведомом самолете с целью измерения всех параметров, необходимых для формирования автоматического управления полетом летательного аппарата. Обработка изображений ведущего самолета выполняется с целью определения координат трех идентичных точек на фоточувствительных матрицах. Причем в качестве этих точек выбираются оптически контрастные элементы конструкции летательного аппарата, например окончания крыла, хвостового оперения и т.д. Для упрощения процедуры обработки изображений возможно использование полупроводниковых источников света в инфракрасном диапазоне (например, с длиной волны λ = 1,54 мкм), что позволяет работать даже в сложных метеоусловиях. Такой подход может быть использован при автоматизации полета строем более чем двух летательных аппаратов, при этом необходимо только оборудовать системой технического зрения все ведомые самолеты группы The article provides a brief analysis of both domestic and foreign inter-aircraft navigation systems. In the course of the analysis, we found the shortcomings of inter-aircraft navigation systems and presented an up-to-date approach to improving the accuracy of the navigation system through the use of a technical vision system. To determine the location of the leading aircraft, we proposed to consider a technical vision system as a measuring complex, which is able to solve a large range of tasks at various stages, in particular, flight in formation. We proposed to install the technical vision system on the slave aircraft in order to measure all the parameters necessary for the formation of automatic flight control of the aircraft. We performed an image processing of the leading aircraft to determine the coordinates of three identical points on photosensitive matrices. Moreover, we selected optically contrasting elements of the aircraft structure as these points, for example, the end of the wing, tail, etc. To simplify the image processing procedure, it is possible to use semiconductor light sources in the infrared range (for example, with a wavelength of λ = 1.54 microns), which allows us to work even in difficult weather conditions. This approach can be used when automating a flight in formation of more than two aircraft, while it is only necessary to equip all the guided aircraft of the group with a technical vision system


2018 ◽  
pp. 458-493
Author(s):  
Li-Minn Ang ◽  
Kah Phooi Seng ◽  
Christopher Wing Hong Ngau

Biological vision components like visual attention (VA) algorithms aim to mimic the mechanism of the human vision system. Often VA algorithms are complex and require high computational and memory requirements to be realized. In biologically-inspired vision and embedded systems, the computational capacity and memory resources are of a primary concern. This paper presents a discussion for implementing VA algorithms in embedded vision systems in a resource constrained environment. The authors survey various types of VA algorithms and identify potential techniques which can be implemented in embedded vision systems. Then, they propose a low complexity and low memory VA model based on a well-established mainstream VA model. The proposed model addresses critical factors in terms of algorithm complexity, memory requirements, computational speed, and salience prediction performance to ensure the reliability of the VA in a resource constrained environment. Finally a custom softcore microprocessor-based hardware implementation on a Field-Programmable Gate Array (FPGA) is used to verify the implementation feasibility of the presented model.


2020 ◽  
Vol 10 (18) ◽  
pp. 6480
Author(s):  
Vicente Román ◽  
Luis Payá ◽  
Sergio Cebollada ◽  
Óscar Reinoso

In this work, an incremental clustering approach to obtain compact hierarchical models of an environment is developed and evaluated. This process is performed using an omnidirectional vision sensor as the only source of information. The method is structured in two loop closure levels. First, the Node Level Loop Closure process selects the candidate nodes with which the new image can close the loop. Second, the Image Level Loop Closure process detects the most similar image and the node with which the current image closed the loop. The algorithm is based on an incremental clustering framework and leads to a topological model where the images of each zone tend to be clustered in different nodes. In addition, the method evaluates when two nodes are similar and they can be merged in a unique node or when a group of connected images are different enough to the others and they should constitute a new node. To perform the process, omnidirectional images are described with global appearance techniques in order to obtain robust descriptors. The use of such technique in mapping and localization algorithms is less extended than local features description, so this work also evaluates the efficiency in clustering and mapping techniques. The proposed framework is tested with three different public datasets, captured by an omnidirectional vision system mounted on a robot while it traversed three different buildings. This framework is able to build the model incrementally, while the robot explores an unknown environment. Some relevant parameters of the algorithm adapt their value as the robot captures new visual information to fully exploit the features’ space, and the model is updated and/or modified as a consequence. The experimental section shows the robustness and efficiency of the method, comparing it with a batch spectral clustering algorithm.


Author(s):  
Xiangyang Xu ◽  
Qiao Chen ◽  
Ruixin Xu

Similar to auditory perception of sound system, color perception of the human visual system also presents a multi-frequency channel property. In order to study the multi-frequency channel mechanism of how the human visual system processes color information, the paper proposed a psychophysical experiment to measure the contrast sensitivities based on 17 color samples of 16 spatial frequencies on CIELAB opponent color space. Correlation analysis was carried out on the psychophysical experiment data, and the results show obvious linear correlations of observations for different spatial frequencies of different observers, which indicates that a linear model can be used to model how human visual system processes spatial frequency information. The results of solving the model based on the experiment data of color samples show that 9 spatial frequency tuning curves can exist in human visual system with each lightness, R–G and Y–B color channel and each channel can be represented by 3 tuning curves, which reflect the “center-around” form of the human visual receptive field. It is concluded that there are 9 spatial frequency channels in human vision system. The low frequency tuning curve of a narrow-frequency bandwidth shows the characteristics of lower level receptive field for human vision system, the medium frequency tuning curve shows a low pass property of the change of medium frequent colors and the high frequency tuning curve of a width-frequency bandwidth, which has a feedback effect on the low and medium frequency channels and shows the characteristics of higher level receptive field for human vision system, which represents the discrimination of details.


2020 ◽  
Vol 2020 (11) ◽  
pp. 234-1-234-6
Author(s):  
Nicolai Behmann ◽  
Holger Blume

LED flicker artefacts, caused by unsynchronized irradiation from a pulse-width modulated LED light source captured by a digital camera sensor with discrete exposure times, place new requirements for both visual and machine vision systems. While latter need to capture relevant information from the light source only in a limited number of frames (e.g. a flickering traffic light), human vision is sensitive to illumination modulation in viewing applications, e.g. digital mirror replacement systems. In order to quantify flicker in viewing applications with KPIs related to human vision, we present a novel approach and results of a psychophysics study on the effect of LED flicker artefacts. Diverse real-world driving sequences have been captured with both mirror replacement cameras and a front viewing camera and potential flicker light sources have been masked manually. Synthetic flicker with adjustable parameters is then overlaid on these areas and the flickering sequences are presented to test persons in a driving environment. Feedback from the testers on flicker perception in different viewing areas, sizes and frequencies are collected and evaluated.


Sign in / Sign up

Export Citation Format

Share Document