scholarly journals Region-Based Segmentation and Object Detection

Object identification and multi-object picture separation are two firmly related processes and it can be enhanced when understood jointly by supporting data from one assignment to the next. Be that as it may, current best in object models are different portrayal for each space creation joint objects and leaving the categorization of numerous part of the scene uncertain. Picture element appearance highlights enable us to do well on classifying formless foundation classes, while the express portrayal of districts encourage the calculation of increasingly complex highlights essential for object detection. Vitally, our model gives a solitary bound together portrayal of the scene we clarify each picture elements of image and authorize it contains in the web between every random variable in our model.

2017 ◽  
Vol 2 (1) ◽  
pp. 80-87
Author(s):  
Puyda V. ◽  
◽  
Stoian. A.

Detecting objects in a video stream is a typical problem in modern computer vision systems that are used in multiple areas. Object detection can be done on both static images and on frames of a video stream. Essentially, object detection means finding color and intensity non-uniformities which can be treated as physical objects. Beside that, the operations of finding coordinates, size and other characteristics of these non-uniformities that can be used to solve other computer vision related problems like object identification can be executed. In this paper, we study three algorithms which can be used to detect objects of different nature and are based on different approaches: detection of color non-uniformities, frame difference and feature detection. As the input data, we use a video stream which is obtained from a video camera or from an mp4 video file. Simulations and testing of the algoritms were done on a universal computer based on an open-source hardware, built on the Broadcom BCM2711, quad-core Cortex-A72 (ARM v8) 64-bit SoC processor with frequency 1,5GHz. The software was created in Visual Studio 2019 using OpenCV 4 on Windows 10 and on a universal computer operated under Linux (Raspbian Buster OS) for an open-source hardware. In the paper, the methods under consideration are compared. The results of the paper can be used in research and development of modern computer vision systems used for different purposes. Keywords: object detection, feature points, keypoints, ORB detector, computer vision, motion detection, HSV model color


i-com ◽  
2008 ◽  
Vol 6 (3/2007) ◽  
pp. 23-29 ◽  
Author(s):  
Birgit Bomsdorf

SummaryTask modelling has entered the development process of web applications, strengthening the usage-centred view within the early steps in Web-Engineering (WE). In current approaches, however, this view is not kept up during subsequent activities to the same degree as this is the case in the field of Human-Computer-Interaction (HCI). The modelling approach presented in this contribution combines models as known from WE with models used in HCI to change this situation. Basically the WE-HCI-integration is supported by combining task and object models as known from HCI with conceptual modelling known from WE. In this paper, the main focus is on the WebTaskModel, a task model adapted to web application concerns, and its contribution towards a task-related web user interface. The main difference to existing task models is the build-time and run-time usage of a generic task lifecycle. Hereby the description of exceptions and erroneous situations during task performance (caused by, e.g., the stateless protocol or Browser interaction) is enabled and at the same time clearly separated from the flow of correct action.


2017 ◽  
Vol 41 (1) ◽  
pp. 94-98 ◽  
Author(s):  
Carolina Mariano Beraldo ◽  
Érika Rondon Lopes ◽  
Raduan Hage ◽  
Maria Cristina F. N. S. Hage

Ingested or penetrating foreign bodies are common in veterinary medicine. When they are radiolucent, these objects become a diagnostic challenge, but they can be investigated sonographically. However, successful object identification depends on the skill of the sonographer. Considering that these cases appear randomly during hospital routines, it is not always possible to train all students to identify them correctly. Therefore, the aim of this study was to produce homemade simulations of radiolucent foreign bodies for veterinary student demonstrations that could be identified sonographically and to evaluate the acceptability, applicability, and usefulness of these simulations according to a visual analog scale questionnaire and subjective questions. For this purpose, object models (a pacifier nipple, a toy ball, a sock, nylon thread, and a mango seed) were designed, produced, and immersed in gelatin. To simulate wood splinters in the integumentary and musculoskeletal system, a piece of meat punctured with a toothpick and ice cream stick splinters were used. The type of phantom had a determinant effect on the visualization (chi-square = 36.528, P < 0.0001) and recognition (chi-square = 18.756, P = 0.0021) capability of the students. All of the students answered that their experience with the models could help in real situations. The student responses to the questionnaire indicated that the project was well accepted, and the participants believed that this experience could be applicable to and useful in veterinary routines.


2020 ◽  
Vol 8 (2) ◽  
pp. 112 ◽  
Author(s):  
Peng Chen ◽  
Ying Li ◽  
Hui Zhou ◽  
Bingxin Liu ◽  
Peng Liu

The synthetic aperture radar (SAR) has a special ability to detect objects in any climate and weather conditions. Consequently, SAR images are widely used in maritime transportation safety and fishery law enforcement for maritime object detection. Currently, deep-learning models are being extensively used for the detection of objects from images. Among them, the feature pyramid network (FPN) uses pyramids for representing semantic information regardless of the scale and has an improved accuracy of object detection. It is also suitable for the detection of multiple small ship objects in SAR images. This study aims to resolve the problems associated with small-object and multi-object ship detection in complex scenarios e.g., when a ship nears the port, by proposing a detection method based on an optimized FPN model. The feature pyramid model is first embedded in a traditional region proposal network (RPN) and mapped into a new feature space for object identification. Subsequently, the k-means clustering algorithm based on the shape similar distance (SSD) measure is used to optimize the FPN. Initial anchor boxes and tests are created using the SAR ship dataset. Experimental results show that the proposed algorithm for object detection shows an accuracy of 98.62%. Compared with Yolo, the RPN based on VGG/ResNet, FPN based on VGG/ResNet, and other models in complex scenarios, the proposed model shows a higher accuracy rate and better overall performance.


2020 ◽  
Vol 2020 (8) ◽  
pp. 221-1-221-7
Author(s):  
Jianhang Chen ◽  
Daniel Mas Montserrat ◽  
Qian Lin ◽  
Edward J. Delp ◽  
Jan P. Allebach

We introduce a new image dataset for object detection and 6D pose estimation, named Extra FAT. The dataset consists of 825K photorealistic RGB images with annotations of groundtruth location and rotation for both the virtual camera and the objects. A registered pixel-level object segmentation mask is also provided for object detection and segmentation tasks. The dataset includes 110 different 3D object models. The object models were rendered in five scenes with diverse illumination, reflection, and occlusion conditions.


Author(s):  
S Gopi Naik

Abstract: The plan is to establish an integrated system that can manage high-quality visual information and also detect weapons quickly and efficiently. It is obtained by integrating ARM-based computer vision and optimization algorithms with deep neural networks able to detect the presence of a threat. The whole system is connected to a Raspberry Pi module, which will capture live broadcasting and evaluate it using a deep convolutional neural network. Due to the intimate interaction between object identification and video and image analysis in real-time objects, By generating sophisticated ensembles that incorporate various low-level picture features with high-level information from object detection and scenario classifiers, their performance can quickly plateau. Deep learning models, which can learn semantic, high-level, deeper features, have been developed to overcome the issues that are present in optimization algorithms. It presents a review of deep learning based object detection frameworks that use Convolutional Neural Network layers for better understanding of object detection. The Mobile-Net SSD model behaves differently in network design, training methods, and optimization functions, among other things. The crime rate in suspicious areas has been reduced as a consequence of weapon detection. However, security is always a major concern in human life. The Raspberry Pi module, or computer vision, has been extensively used in the detection and monitoring of weapons. Due to the growing rate of human safety protection, privacy and the integration of live broadcasting systems which can detect and analyse images, suspicious areas are becoming indispensable in intelligence. This process uses a Mobile-Net SSD algorithm to achieve automatic weapons and object detection. Keywords: Computer Vision, Weapon and Object Detection, Raspberry Pi Camera, RTSP, SMTP, Mobile-Net SSD, CNN, Artificial Intelligence.


Author(s):  
D.P. Tripathy ◽  
K. Guru Raghavendra Reddy

Moving object detection is an important task in many computer vision classifications applications. The goal of this study is to identify a moving object detection method that provides a reliable and accurate identification of objects on the conveyor belt. In this paper, a study of the moving object detection methods is presented. Firstly, moving object detection pixel by pixel was performed using background subtraction, frame difference method. The threshold value in both background subtraction and frame difference is a fixed value, which determines the accuracy of object identification. The adaptive threshold values were calculated for both the methods to improve the accuracy. The performance of these methods was compared with the ground truth image.


2020 ◽  
Vol 17 (1) ◽  
pp. 439-444 ◽  
Author(s):  
Katamneni Vinaya Sree ◽  
G. Jeyakumar

In the given image identifying the existence of a required object is the concern of the object detection process. This is quite natural for Human without any effort, however making a machine to detect an object in image is tedious. To make machines to recognize the objects, the feature descriptor algorithms are to be implemented. The general object detection approaches use collection of local and global descriptors to represent an image. Difficulties arise during this process when there is variation in lightening, positioning, rotation, mirroring, occlusion, scaling etc., of the same object in different image scenes. To overcome these difficulties, we need combination of features that detects the object in the image scene. But there exist lot of descriptors that can be used. Hence, finding the required number of feature descriptors for object detection is a crucial task. The question that comes out here is how to select the optimum number of descriptors to achieve optimum accuracy? The answer for the question is an optimization algorithm, which can be employed to select the best combination of the descriptors with maximum detection accuracy. This paper proposing an Evolutionary Computation (EC) based approach with the Differential Evolution (DE) algorithm to find the optimal combination of feature descriptors to achieve optimal object detection accuracy. The proposed approach is implemented and its superiority is verified with four different images and results obtained are presented in this paper.


2021 ◽  
Vol 13 (12) ◽  
pp. 307
Author(s):  
Vijayakumar Varadarajan ◽  
Dweepna Garg ◽  
Ketan Kotecha

Deep learning is a relatively new branch of machine learning in which computers are taught to recognize patterns in massive volumes of data. It primarily describes learning at various levels of representation, which aids in understanding data that includes text, voice, and visuals. Convolutional neural networks have been used to solve challenges in computer vision, including object identification, image classification, semantic segmentation and a lot more. Object detection in videos involves confirming the presence of the object in the image or video and then locating it accurately for recognition. In the video, modelling techniques suffer from high computation and memory costs, which may decrease performance measures such as accuracy and efficiency to identify the object accurately in real-time. The current object detection technique based on a deep convolution neural network requires executing multilevel convolution and pooling operations on the entire image to extract deep semantic properties from it. For large objects, detection models can provide superior results; however, those models fail to detect the varying size of the objects that have low resolution and are greatly influenced by noise because the features after the repeated convolution operations of existing models do not fully represent the essential characteristics of the objects in real-time. With the help of a multi-scale anchor box, the proposed approach reported in this paper enhances the detection accuracy by extracting features at multiple convolution levels of the object. The major contribution of this paper is to design a model to understand better the parameters and the hyper-parameters which affect the detection and the recognition of objects of varying sizes and shapes, and to achieve real-time object detection and recognition speeds by improving accuracy. The proposed model has achieved 84.49 mAP on the test set of the Pascal VOC-2007 dataset at 11 FPS, which is comparatively better than other real-time object detection models.


Sign in / Sign up

Export Citation Format

Share Document