scholarly journals Perception and prediction of the putting distance of robot putting movements under different visual/viewing conditions

PLoS ONE ◽  
2021 ◽  
Vol 16 (4) ◽  
pp. e0249518
Author(s):  
Gerrit Kollegger ◽  
Josef Wiemeyer ◽  
Marco Ewerton ◽  
Jan Peters

The purpose of this paper is to examine, whether and under which conditions humans are able to predict the putting distance of a robotic device. Based on the “flash-lag effect” (FLE) it was expected that the prediction errors increase with increasing putting velocity. Furthermore, we hypothesized that the predictions are more accurate and more confident if human observers operate under full vision (F-RCHB) compared to either temporal occlusion (I-RCHB) or spatial occlusion (invisible ball, F-RHC, or club, F-B). In two experiments, 48 video sequences of putt movements performed by a BioRob robot arm were presented to thirty-nine students (age: 24.49±3.20 years). In the experiments, video sequences included six putting distances (1.5, 2.0, 2.5, 3.0, 3.5, and 4.0 m; experiment 1) under full versus incomplete vision (F-RCHB versus I-RCHB) and three putting distances (2. 0, 3.0, and 4.0 m; experiment 2) under the four visual conditions (F-RCHB, I-RCHB, F-RCH, and F-B). After the presentation of each video sequence, the participants estimated the putting distance on a scale from 0 to 6 m and provided their confidence of prediction on a 5-point scale. Both experiments show comparable results for the respective dependent variables (error and confidence measures). The participants consistently overestimated the putting distance under the full vision conditions; however, the experiments did not show a pattern that was consistent with the FLE. Under the temporal occlusion condition, a prediction was not possible; rather a random estimation pattern was found around the centre of the prediction scale (3 m). Spatial occlusion did not affect errors and confidence of prediction. The experiments indicate that temporal constraints seem to be more critical than spatial constraints. The FLE may not apply to distance prediction compared to location estimation.

2011 ◽  
Vol 225-226 ◽  
pp. 403-406
Author(s):  
Xin Zhang ◽  
Xiao Tao Wang ◽  
Bing Wang ◽  
Yue Hua Gao

Human Skin Color(HSC) features have been widely used in video moving human positioning. However, in complex background video sequences, due to illumination changes or other moving objects which have similar HSC regions, the effect of moving human positioning is not satisfactory. A new method of moving human positioning applied on complex background video sequences is presented in this paper. Firstly, brightness information of the video sequence images is detected and analyzed based on HSV color model. Secondly, adopt the multi frame subtraction method to extract the moving object regions from motionless background. Then, the regions with distinctive HSC features are separated from other moving objects using the data fusion model of HSC and brightness information. Finally, identify human object among regions with HSC features according to the prior knowledge of human. The experimental results show that the method provided in this paper is effective in moving human positioning of complex background video, and has the strong illumination change adaptability and anti-jamming ability.


Sensors ◽  
2020 ◽  
Vol 20 (18) ◽  
pp. 5368
Author(s):  
Atul Sharma ◽  
Sushil Raut ◽  
Kohei Shimasaki ◽  
Taku Senoo ◽  
Idaku Ishii

This study develops a projector–camera-based visible light communication (VLC) system for real-time broadband video streaming, in which a high frame rate (HFR) projector can encode and project a color input video sequence into binary image patterns modulated at thousands of frames per second and an HFR vision system can capture and decode these binary patterns into the input color video sequence with real-time video processing. For maximum utilization of the high-throughput transmission ability of the HFR projector, we introduce a projector–camera VLC protocol, wherein a multi-level color video sequence is binary-modulated with a gray code for encoding and decoding instead of pure-code-based binary modulation. Gray code encoding is introduced to address the ambiguity with mismatched pixel alignments along the gradients between the projector and vision system. Our proposed VLC system consists of an HFR projector, which can project 590 × 1060 binary images at 1041 fps via HDMI streaming and a monochrome HFR camera system, which can capture and process 12-bit 512 × 512 images in real time at 3125 fps; it can simultaneously decode and reconstruct 24-bit RGB video sequences at 31 fps, including an error correction process. The effectiveness of the proposed VLC system was verified via several experiments by streaming offline and live video sequences.


2013 ◽  
Vol 479-480 ◽  
pp. 174-178
Author(s):  
Shi Wei Lo

This paper addresses a compact framework to matching video sequences through a PSNR-based profile. This simplify video profile is suitable to matching process when apply in disordered undersea videos. As opposed to using color and motion feature across the video sequence, we use the image quality of successive frames to be a feature of videos. We employ the PSNR quality feature to be a video profile rather than the complex contend-based analysis. The experimental results show that the proposed approach permits accurate of matching video. The performance is satisfactory on determine correct video from undersea dataset.


2012 ◽  
Vol 21 (01) ◽  
pp. 1250012 ◽  
Author(s):  
HUCHUAN LU ◽  
SHIPENG LU ◽  
GANG YANG

In this paper, we present a novel method for eye tracking, in detail describing the eye contour and the visible iris center. Combining the IVT (Incremental Visual Tracking) tracker, the proposed online affine manifold model, in which the sequentially learning shape and texture are modeled in the first stage and noniterative recovering estimation in the second stage, tracks the eye contour in video sequences. After that, an adaptive black round mask is generated to match the visible iris center. Experimental results of eye tracking indicate that our tracker works well in the PC or domestic camera captured image streams with considerable head and eyeball rotation.


2012 ◽  
Vol 457-458 ◽  
pp. 867-871
Author(s):  
Xi Nan Zhang ◽  
Ai Lin Liu

To reduce the complexity of AVS sub-pixel motion vector search, this paper proposes a new improved algorithm based on little diamond window searching policies according to the law of the general video sequence motion vector focusing on the near of the initial searching point at the time of the subpixel motion estimation on the basis of deeping analysis to HFPS algorithm. Comparing with the sub-pixel HFPS algorithm search algorithm, the time of sub-pixel motion estimation can reduce 51.97% on average and efficiently decrease the computational number of sub-pixel motion estimation on average PSNR lost less than 0.01dB to the video sequences with different motion characteristics.


2013 ◽  
Vol 347-350 ◽  
pp. 3500-3504
Author(s):  
Xiao Ran Guo ◽  
Shao Hui Cui ◽  
Fang Dan

This article presents a novel approach to extract robust local feature points of video sequence in digital image stabilization system. Robust Harris-SIFT detector is proposed to select the most stable SIFT key points in the video sequence where image motion is happened due to vehicle or platform vibration. Experimental results show that the proposed scheme is robust to various transformations of video sequences, such as translation, rotation and scaling, as well as blurring. Compared with the current state-of-the-art schemes, the proposed scheme yields better performances.


Author(s):  
Luis Carlos Manrique ◽  
Anthony Weiss ◽  
Sandra Puentes

The use of electronic devices designed for user location estimation has become widely popular in the last decade. This is thanks to emergent technologies such as Bluetooth Low Energy, Radio-Frequency Identification, and Ultra-WideBand (UWB) among others. In the present study; the authors provide a method for using a Geographic Information System (GIS) to define spatial constraints, in order to simulate the lines of sight of anchors to make an informed selection of adequate locations for installation. By leveraging GIS, researchers or enterprises can improve the installation process by reducing costs while setting up arrangements that will ensure reliable data collection. We include a scenario illustrating the possibility of budget reduction of around 30% related to the orientation and survey of the devices.


Author(s):  
Boris Faizov ◽  
Vlad Shakhuro ◽  
Anton Konushin

This work is devoted to the automatic detection of unwanted driver behavior such as smoking, using a mobile phone, and eating. The various existing datasets are practically unsuitable for this task. We did not find suitable training data with RGB video sequences shot from the position of the inner mirror. So we investigated the possibility of training the algorithms for this task on an out-of-domain set of people faces images. We also filmed our own test video sequence in a car to test the algorithms. We investigated different existing algorithms working both with one frame and with video sequences and conducted an experimental comparison of them. The availability of temporal information improved quality. Another important aspect is metrics for assessing the quality of the resulting system. We showed that experimental evaluation in this task should be performed on the entire video sequences. We proposed an algorithm for detecting undesirable driver actions and showed its effectiveness.


Fractals ◽  
1994 ◽  
Vol 02 (03) ◽  
pp. 391-394 ◽  
Author(s):  
DONALD M MONRO ◽  
JEREMY A NICHOLLS

Using a first order (bilinear) Bath Fractal Transform (BFT), we have obtained useful video sequences of talking heads with transmission rates as low as 40 KBits/sec. Previously, fractal image coding had been computationally asymmetric. In our work, 8 by 8 pixel blocks are coded in 250 µs and decoded in 200 µs on a 33 MHz i-486 based PC. This is of significance in consumer electronics, such as personal communications, where inexpensive coding systems will have an advantage over more expensive methods using DSP or custom chips to achieve the necessary speed. With a simple quantization and entropy coding scheme applied to standard QCIF frames, at 40 KBits/sec we achieve coding of 40% of each frame in a 25 Hz video sequence, equivalent to 100% at 10 Hz.


Author(s):  
D. Murashov ◽  
Y. Obukhov ◽  
I. Kershner ◽  
M. Sinkin

<p><strong>Abstract.</strong> In this paper, an algorithm for automated detecting diagnostic events in video channel of video and electroencephalographic (EEG) monitoring data is presented. The analysis of video sequences is focused on identifying a group of frames with high or very low (depending on the type of seizure) dynamics of informative areas according to a criterion calculated during processing of the optical flow. The preliminary results of the analysis of real clinical data are given and compared with data obtained from the synchronous EEG. The results showed the possibility in principle of reliable diagnosing epileptic seizures and distinguishing them from non-epileptic events.</p>


Sign in / Sign up

Export Citation Format

Share Document