Android Based Object Detection System for Visually Impaired

Author(s):  
Ajinkya Badave ◽  
Rathin Jagtap ◽  
Rizina Kaovasia ◽  
Shivani Rahatwad ◽  
Saroja Kulkarni
2019 ◽  
Vol 8 (2S8) ◽  
pp. 1675-1676

The point of this paper is to research the improvement of a route help for visually impaired and outwardly weakened People. It has microcontroller which has wifi inbuilt module. This guide is convenient and offers data to the client to move around in new condition, regardless of whether indoor or open air, through an easy to use interface. Then again, and so as to lessen route challenges of the visually impaired, a deterrent location framework utilizing ultrasounds and vibrators is added to this gadget. The proposed framework identifies the closest hindrance through ultrasonic sensors and it gives an alert to illuminate the visually impaired about its confinement.


Author(s):  
Raghad Raied Mahmood Et al.

It is relatively simple for a normal human to interpret and understand every banknote, but one of the major problems for visually impaired people are money recognition, especially for paper currency. Since money plays such an important role in our everyday lives and is required for every business transaction, real-time detection and recognition of banknotes become a necessity for blind or visually impaired people For that purpose, we propose a real-time object detection system to help visually impaired people in their daily business transactions. Dataset Images of the Iraqi banknote category are collected in different conditions initially and then, these images are augmented with different geometric transformations, to make the system strong. These augmented images are then annotated manually using the "LabelImg" program, from which training sets and validation image sets are prepared. We will use YOLOv3 real-time Object Detection algorithm trained on custom Iraqi banknote dataset for detection and recognition of banknotes. Then the label of the banknotes is identified and then converted into audio by using Google Text to Speech (gTTS), which will be the expected output. The performance of the trained model is evaluated on a test dataset and real-time live video. The test results demonstrate that the proposed method can detect and recognize Iraqi paper money with high mAP reaches 97.405% and a short time.


Author(s):  
Rajeshvaree Ravindra Karmarkar ◽  
Prof.V.N Honmane

—As object recognition technology has developed recently, various technologies have been applied to autonomousvehicles, robots, and industrial facilities. However, the benefits ofthese technologies are not reaching the visually impaired, who need it the most. This paper proposed an object detection system for the blind using deep learning technologies. Furthermore, a voice guidance technique is used to inform sight impaired persons as to the location of objects. The object recognition deep learning model utilizes the You Only Look Once(YOLO) algorithm and a voice announcement is synthesized using text-tospeech (TTS) to make it easier for the blind to get information about objects. Asa result, it implements an efficient object-detection system that helps the blind find objects in a specific space without help from others, and the system is analyzed through experiments to verify performance.


2021 ◽  
pp. 1631-1642
Author(s):  
D. Ravi Kumar ◽  
Hiren Kumar Thakkar ◽  
Suresh Merugu ◽  
Vinit Kumar Gunjan ◽  
Suneet K. Gupta

It is realized that innovative progressions are expanding at a quicker pace. In any case, the usage of these innovations is less in different parts. It is realized that the general population of nowadays need help for doing a few works when they were matured. .It is troublesome for visually impaired individuals to recognize these articles whenever lostinside a home. But since of these actualities, we may encounter different pressure and issues identified with the loss of those items for individuals. So, we propose a framework where we candistinguishthelostarticleswiththeassistanceofourproposedfram eworkutilizingImagehandlingprocedures.Thatadditionallyreadyto supporttheindividualdependentontheface recognized and confine or personal the passage of obscure people. The current frameworkdistinguishesthelostarticleswiththeassistanceofGPS.Ye t,theuseoffaceacknowledgement for individual recognizable proof by robots is not being used. It doesn't give proficient yield. This causes different slack in the circuit. So we propose a framework to upgrade the item following.


2021 ◽  
Vol 11 (11) ◽  
pp. 4894
Author(s):  
Anna Scius-Bertrand ◽  
Michael Jungo ◽  
Beat Wolf ◽  
Andreas Fischer ◽  
Marc Bui

The current state of the art for automatic transcription of historical manuscripts is typically limited by the requirement of human-annotated learning samples, which are are necessary to train specific machine learning models for specific languages and scripts. Transcription alignment is a simpler task that aims to find a correspondence between text in the scanned image and its existing Unicode counterpart, a correspondence which can then be used as training data. The alignment task can be approached with heuristic methods dedicated to certain types of manuscripts, or with weakly trained systems reducing the required amount of annotations. In this article, we propose a novel learning-based alignment method based on fully convolutional object detection that does not require any human annotation at all. Instead, the object detection system is initially trained on synthetic printed pages using a font and then adapted to the real manuscripts by means of self-training. On a dataset of historical Vietnamese handwriting, we demonstrate the feasibility of annotation-free alignment as well as the positive impact of self-training on the character detection accuracy, reaching a detection accuracy of 96.4% with a YOLOv5m model without using any human annotation.


Sensors ◽  
2021 ◽  
Vol 21 (16) ◽  
pp. 5279
Author(s):  
Dong-Hoon Kwak ◽  
Guk-Jin Son ◽  
Mi-Kyung Park ◽  
Young-Duk Kim

The consumption of seaweed is increasing year by year worldwide. Therefore, the foreign object inspection of seaweed is becoming increasingly important. Seaweed is mixed with various materials such as laver and sargassum fusiforme. So it has various colors even in the same seaweed. In addition, the surface is uneven and greasy, causing diffuse reflections frequently. For these reasons, it is difficult to detect foreign objects in seaweed, so the accuracy of conventional foreign object detectors used in real manufacturing sites is less than 80%. Supporting real-time inspection should also be considered when inspecting foreign objects. Since seaweed requires mass production, rapid inspection is essential. However, hyperspectral imaging techniques are generally not suitable for high-speed inspection. In this study, we overcome this limitation by using dimensionality reduction and using simplified operations. For accuracy improvement, the proposed algorithm is carried out in 2 stages. Firstly, the subtraction method is used to clearly distinguish seaweed and conveyor belts, and also detect some relatively easy to detect foreign objects. Secondly, a standardization inspection is performed based on the result of the subtraction method. During this process, the proposed scheme adopts simplified and burdenless calculations such as subtraction, division, and one-by-one matching, which achieves both accuracy and low latency performance. In the experiment to evaluate the performance, 60 normal seaweeds and 60 seaweeds containing foreign objects were used, and the accuracy of the proposed algorithm is 95%. Finally, by implementing the proposed algorithm as a foreign object detection platform, it was confirmed that real-time operation in rapid inspection was possible, and the possibility of deployment in real manufacturing sites was confirmed.


Sign in / Sign up

Export Citation Format

Share Document