scholarly journals Computer Vision and Deep Learning for Environment-Adaptive Control of Robotic Lower-Limb Exoskeletons

2021 ◽  
Author(s):  
Brokoslaw Laschowski ◽  
William McNally ◽  
Alexander Wong ◽  
John McPhee

Robotic exoskeletons require human control and decision making to switch between different locomotion modes, which can be inconvenient and cognitively demanding. To support the development of automated locomotion mode recognition systems (i.e., high-level controllers), we designed an environment recognition system using computer vision and deep learning. We collected over 5.6 million images of indoor and outdoor real-world walking environments using a wearable camera system, of which ~923,000 images were annotated using a 12-class hierarchical labelling architecture (called the ExoNet database). We then trained and tested the EfficientNetB0 convolutional neural network, designed for efficiency using neural architecture search, to predict the different walking environments. Our environment recognition system achieved ~73% image classification accuracy. While these preliminary results benchmark EfficientNetB0 on the ExoNet database, further research is needed to compare different image classification algorithms to develop an accurate and real-time environment-adaptive locomotion mode recognition system for robotic exoskeleton control.

Sensors ◽  
2021 ◽  
Vol 21 (2) ◽  
pp. 526
Author(s):  
Yang Han ◽  
Chunbao Liu ◽  
Lingyun Yan ◽  
Lei Ren

Smart wearable robotic system, such as exoskeleton assist device and powered lower limb prostheses can rapidly and accurately realize man–machine interaction through locomotion mode recognition system. However, previous locomotion mode recognition studies usually adopted more sensors for higher accuracy and effective intelligent algorithms to recognize multiple locomotion modes simultaneously. To reduce the burden of sensors on users and recognize more locomotion modes, we design a novel decision tree structure (DTS) based on using an improved backpropagation neural network (IBPNN) as judgment nodes named IBPNN-DTS, after analyzing the experimental locomotion mode data using the original values with a 200-ms time window for a single inertial measurement unit to hierarchically identify nine common locomotion modes (level walking at three kinds of speeds, ramp ascent/descent, stair ascent/descent, Sit, and Stand). In addition, we reduce the number of parameters in the IBPNN for structure optimization and adopted the artificial bee colony (ABC) algorithm to perform global search for initial weight and threshold value to eliminate system uncertainty because randomly generated initial values tend to result in a failure to converge or falling into local optima. Experimental results demonstrate that recognition accuracy of the IBPNN-DTS with ABC optimization (ABC-IBPNN-DTS) was up to 96.71% (97.29% for the IBPNN-DTS). Compared to IBPNN-DTS without optimization, the number of parameters in ABC-IBPNN-DTS shrank by 66% with only a 0.58% reduction in accuracy while the classification model kept high robustness.


2018 ◽  
Vol 7 (2.7) ◽  
pp. 614 ◽  
Author(s):  
M Manoj krishna ◽  
M Neelima ◽  
M Harshali ◽  
M Venu Gopala Rao

The image classification is a classical problem of image processing, computer vision and machine learning fields. In this paper we study the image classification using deep learning. We use AlexNet architecture with convolutional neural networks for this purpose. Four test images are selected from the ImageNet database for the classification purpose. We cropped the images for various portion areas and conducted experiments. The results show the effectiveness of deep learning based image classification using AlexNet.  


2020 ◽  
Vol 61 (82) ◽  
pp. 127-138
Author(s):  
Scott Sorensen ◽  
Vinit Veerendraveer ◽  
Wayne Treible ◽  
Andrew R. Mahoney ◽  
Chandra Kambhamettu

AbstractThe Polar Sea Ice Topography REconstruction System, or PSITRES, is a 3D camera system designed to continuously monitor an area of ice and water adjacent to an ice-going vessel. Camera systems aboard ships in the polar regions are common; however, the application of computer vision techniques to extract high-level information from the imagery is infrequent. Many of the existing systems are built for human involvement throughout the process and lack automation necessary for round the clock use. The PSITRES was designed with computer vision in mind. It can capture images continuously for days on end with limited oversight. We have applied the system in different ice observing scenarios. The PSITRES was deployed on three research expeditions in the Arctic and Subarctic, and we present applications in measuring ice concentration, melt pond fraction and presence of algae. Systems like PSITRES and the computer vision algorithms applied represent steps toward automatically observing, evaluating and analyzing ice and the environment around ships in ice-covered waters.


2022 ◽  
Vol 2022 ◽  
pp. 1-8
Author(s):  
Xianben Yang ◽  
Wei Zhang

In recent years, due to the wide application of deep learning and more modal research, the corresponding image retrieval system has gradually extended from traditional text retrieval to visual retrieval combined with images and has become the field of computer vision and natural language understanding and one of the important cross-research hotspots. This paper focuses on the research of graph convolutional networks for cross-modal information retrieval and has a general understanding of cross-modal information retrieval and the related theories of convolutional networks on the basis of literature data. Modal information retrieval is designed to combine high-level semantics with low-level visual capabilities in cross-modal information retrieval to improve the accuracy of information retrieval and then use experiments to verify the designed network model, and the result is that the model designed in this paper is more accurate than the traditional retrieval model, which is up to 90%.


Author(s):  
S Gopi Naik

Abstract: The plan is to establish an integrated system that can manage high-quality visual information and also detect weapons quickly and efficiently. It is obtained by integrating ARM-based computer vision and optimization algorithms with deep neural networks able to detect the presence of a threat. The whole system is connected to a Raspberry Pi module, which will capture live broadcasting and evaluate it using a deep convolutional neural network. Due to the intimate interaction between object identification and video and image analysis in real-time objects, By generating sophisticated ensembles that incorporate various low-level picture features with high-level information from object detection and scenario classifiers, their performance can quickly plateau. Deep learning models, which can learn semantic, high-level, deeper features, have been developed to overcome the issues that are present in optimization algorithms. It presents a review of deep learning based object detection frameworks that use Convolutional Neural Network layers for better understanding of object detection. The Mobile-Net SSD model behaves differently in network design, training methods, and optimization functions, among other things. The crime rate in suspicious areas has been reduced as a consequence of weapon detection. However, security is always a major concern in human life. The Raspberry Pi module, or computer vision, has been extensively used in the detection and monitoring of weapons. Due to the growing rate of human safety protection, privacy and the integration of live broadcasting systems which can detect and analyse images, suspicious areas are becoming indispensable in intelligence. This process uses a Mobile-Net SSD algorithm to achieve automatic weapons and object detection. Keywords: Computer Vision, Weapon and Object Detection, Raspberry Pi Camera, RTSP, SMTP, Mobile-Net SSD, CNN, Artificial Intelligence.


Sensors ◽  
2021 ◽  
Vol 21 (2) ◽  
pp. 327
Author(s):  
Ramiz Yilmazer ◽  
Derya Birant

Providing high on-shelf availability (OSA) is a key factor to increase profits in grocery stores. Recently, there has been growing interest in computer vision approaches to monitor OSA. However, the largest and well-known computer vision datasets do not provide annotation for store products, and therefore, a huge effort is needed to manually label products on images. To tackle the annotation problem, this paper proposes a new method that combines two concepts “semi-supervised learning” and “on-shelf availability” (SOSA) for the first time. Moreover, it is the first time that “You Only Look Once” (YOLOv4) deep learning architecture is used to monitor OSA. Furthermore, this paper provides the first demonstration of explainable artificial intelligence (XAI) on OSA. It presents a new software application, called SOSA XAI, with its capabilities and advantages. In the experimental studies, the effectiveness of the proposed SOSA method was verified on image datasets, with different ratios of labeled samples varying from 20% to 80%. The experimental results show that the proposed approach outperforms the existing approaches (RetinaNet and YOLOv3) in terms of accuracy.


Medical imaging classification is playing a vital role in identifying and diagnoses the diseases, which is very helpful to doctor. Conventional ways classify supported the form, color, and/or texture, most of tiny problematic areas haven’t shown in medical images, which meant less efficient classification and that has poor ability to identify disease. Advanced deep learning algorithms provide an efficient way to construct a finished model that can compute final classification labels with the raw pixels of medical images. These conventional algorithms are not sufficient for high resolution images due to small dataset size, advanced deep learning models suffer from very high computational costs and limitations in the channels and multilayers in the channels. To overcome these limitations, we proposed a new algorithm Normalized Coding Network with Multi-scale Perceptron (NCNMP), which combines high-level features and traditional features. The Architecture of the proposed model includes three stages. Training, retrieve, fuse. We examined the proposed algorithm on medical image dataset NIH2626. We got an overall image classification accuracy of 91.35, which are greater than the present methods.


Sebatik ◽  
2020 ◽  
Vol 24 (2) ◽  
pp. 300-306
Author(s):  
Muhamad Jaelani Akbar ◽  
Mochamad Wisuda Sardjono ◽  
Margi Cahyanti ◽  
Ericks Rachmat Swedia

Sayuran merupakan sebutan bagi bahan pangan asal tumbuhan yang biasanya mengandung kadar air tinggi dan dikonsumsi dalam keadaan segar atau setelah diolah secara minimal. Keanekaragaman sayur yang terdapat di dunia menyebabkan keragaman pula dalam pengklasifikasian sayur. Oleh karena itu diperlukan adanya pendekatan digital agar dapat mengenali jenis sayuran dengan cepat dan mudah. Dalam penelitian ini jumlah jenis sayuran yang digunakan sebanyak 7 jenis diantara: brokoli, jagung, kacang panjang, pare, terung ungu, tomat dan kubis. Dataset yang digunakan berjumlah 941 gambar sayur dari 7 jenis sayur, ditambah 131 gambar sayur dari jenis yang tidak terdapat pada dataset, selain itu digunakan 291 gambar selain sayuran. Untuk melakukan klasifikasi jenis sayuran digunakan algoritme Convolutional Neural Network (CNN), yang merupakan salah satu bidang ilmu baru dalam Machine Learning dan berkembang dengan pesat. CNN merupakan salah satu algoritme yang terdapat pada metode Deep Learning dengan memiliki kemampuan yang baik dalam Computer Vision, salah satunya yaitu image classification atau klasifikasi objek citra. Uji coba dilakukan pada lima perangkat selular berbasiskan sistem operasi Android. Python digunakan sebagai bahasa pemrograman dalam merancang aplikasi mobile ini dengan menggunakan modul Tensor flow untuk melakukan training dan testing data. Metode yang dapat digunakan dalam melakukan klasifikasi citra ini yaitu Convolutional Neural Network (CNN). Hasil final test accuracy yang diperoleh yaitu didapat keakuratan mengenali jenis sayuran sebesar 98.1% dengan salah satu hasil pengujian yaitu klasifikasi sayur jagung dengan akurasi sebesar 99.98049%.


2021 ◽  
Author(s):  
Daniel Padilla ◽  
Hatem A. Rashwan ◽  
Domènec Savi Puig

Deep learning (DL) networks have proven to be crucial in commercial solutions with computer vision challenges due to their abilities to extract high-level abstractions of the image data and their capabilities of being easily adapted to many applications. As a result, DL methodologies had become a de facto standard for computer vision problems yielding many new kinds of research, approaches and applications. Recently, the commercial sector is also driving to use of embedded systems to be able to execute DL models, which has caused an important change on the DL panorama and the embedded systems themselves. Consequently, in this paper, we attempt to study the state of the art of embedded systems, such as GPUs, FPGAs and Mobile SoCs, that are able to use DL techniques, to modernize the stakeholders with the new systems available in the market. Besides, we aim at helping them to determine which of these systems can be beneficial and suitable for their applications in terms of upgradeability, price, deployment and performance.


Sign in / Sign up

Export Citation Format

Share Document