scholarly journals A Deep Dive of Autoencoder Models on Low-Contrast Aquatic Images

Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 4966
Author(s):  
Rich C. Lee ◽  
Ing-Yi Chen

Public aquariums and similar institutions often use video as a method to monitor the behavior, health, and status of aquatic organisms in their environments. These video footages take up a sizeable amount of space and require the use of autoencoders to reduce their file size for efficient storage. The autoencoder neural network is an emerging technique which uses the extracted latent space from an input source to reduce the image size for storage, and then reconstructs the source within an acceptable loss range for use. To meet an aquarium’s practical needs, the autoencoder must have easily maintainable codes, low power consumption, be easily adoptable, and not require a substantial amount of memory use or processing power. Conventional configurations of autoencoders often provide results that perform beyond an aquarium’s needs at the cost of being too complex for their architecture to handle, while few take low-contrast sources into consideration. Thus, in this instance, “keeping it simple” would be the ideal approach to the autoencoder’s model design. This paper proposes a practical approach catered to an aquarium’s specific needs through the configuration of autoencoder parameters. It first explores the differences between the two of the most widely applied autoencoder approaches, Multilayer Perceptron (MLP) and Convolution Neural Networks (CNN), to identify the most appropriate approach. The paper concludes that while both approaches (with proper configurations and image preprocessing) can reduce the dimensionality and reduce visual noise of the low-contrast images gathered from aquatic video footage, the CNN approach is more suitable for an aquarium’s architecture. As an unexpected finding of the experiments conducted, the paper also discovered that by manipulating the formula for the MLP approach, the autoencoder could generate a denoised differential image that contains sharper and more desirable visual information to an aquarium’s operation. Lastly, the paper has found that proper image preprocessing prior to the application of the autoencoder led to better model convergence and prediction results, as demonstrated both visually and numerically in the experiment. The paper concludes that by combining the denoising effect of MLP, CNN’s ability to manage memory consumption, and proper image preprocessing, the specific practical needs of an aquarium can be adeptly fulfilled.

2019 ◽  
Author(s):  
Meghana Srivatsav ◽  
Timothy John Luke ◽  
Pär Anders Granhag ◽  
Leif Strömwall ◽  
Aldert Vrij

With Study 1 (N=140), we aimed to examine how different ways of disclosing evidence during an interview would influence guilty suspects’ perception of interviewer’s prior knowledge and elicit statement-evidence inconsistencies. We predicted that interviews with evidence disclosed would elicit low statement-evidence inconsistencies whereas interviews where evidence was not disclosed would result in high statement-evidence inconsistencies. The outcome did not support our predictions. Guilty suspects revealed crime-related information about non-critical themes and withheld information regarding critical themes irrespective of evidence disclosure. We explored this unexpected finding in Study 2 (N=216), which was designed to understand if guilty suspects would reveal information regarding themes of the crime that are not incriminating (not critical) in comparison to themes that were incriminating (critical) as observed in Study 1. We used the evidence disclosure tactics of Study 1 in Study 2 and also measured how these influence their perception of interviewer’s knowledge. The outcome replicated findings from Study 1 that guilty suspects reveal or withhold information based on the cost of disclosing the information. This is a novel finding in the Strategic Use of Evidence literature.


2021 ◽  
Author(s):  
Ying Bi ◽  
Bing Xue ◽  
Mengjie Zhang

Being able to learn discriminative features from low-quality images has raised much attention recently due to their wide applications ranging from autonomous driving to safety surveillance. However, this task is difficult due to high variations across images, such as scale, rotation, illumination, and viewpoint, and distortions in images, such as blur, low contrast, and noise. Image preprocessing could improve the quality of the images, but it often requires human intervention and domain knowledge. Genetic programming (GP) with a flexible representation can automatically perform image preprocessing and feature extraction without human intervention. Therefore, this study proposes a new evolutionary learning approach using GP (EFLGP) to learn discriminative features from images with blur, low contrast, and noise for classification. In the proposed approach, we develop a new program structure (individual representation), a new function set, and a new terminal set. With these new designs, EFLGP can detect small regions from a large input low-quality image, select image operators to process the regions or detect features from the small regions, and output a flexible number of discriminative features. A set of commonly used image preprocessing operators is employed as functions in EFLGP to allow it to search for solutions that can effectively handle low-quality image data. The performance of EFLGP is comprehensively investigated on eight datasets of varying difficulty under the original (clean), blur, low contrast, and noise scenarios, and compared with a large number of benchmark methods using handcrafted features and deep features. The experimental results show that EFLGP achieves significantly better or similar results in most comparisons. The results also reveal that EFLGP is more invariant than the benchmark methods to blur, low contrast, and noise.


Author(s):  
Shilin Wang ◽  
Wing Hong Lau ◽  
Alan Wee-Chung Liew ◽  
Shu Hung Leung

Recently, lip image analysis has received much attention because the visual information extracted has been shown to provide significant improvement for speech recognition and speaker authentication, especially in noisy environments. Lip image segmentation plays an important role in lip image analysis. This chapter will describe different lip image segmentation techniques, with emphasis on segmenting color lip images. In addition to providing a review of different approaches, we will describe in detail the state-of-the-art classification-based techniques recently proposed by our group for color lip segmentation: “Spatial fuzzy c-mean clustering” (SFCM) and “fuzzy c-means with shape function” (FCMS). These methods integrate the color information along with different kinds of spatial information into a fuzzy clustering structure and demonstrate superiority in segmenting color lip images with natural low contrast in comparison with many traditional image segmentation techniques.


Symmetry ◽  
2018 ◽  
Vol 10 (10) ◽  
pp. 491 ◽  
Author(s):  
Lincheng Jiang ◽  
Yumei Jing ◽  
Shengze Hu ◽  
Bin Ge ◽  
Weidong Xiao

Due to the cost limitation of camera sensors, images captured in low-light environments often suffer from low contrast and multiple types of noise. A number of algorithms have been proposed to improve contrast and suppress noise in the input low-light images. In this paper, a deep refinement network, LL-RefineNet, is built to learn from the synthetical dark and noisy training images, and perform image enhancement for natural low-light images in symmetric—forward and backward—pathways. The proposed network utilizes all the useful information from the down-sampling path to produce the high-resolution enhancement result, where global features captured from deeper layers are gradually refined using local features generated by earlier convolutions. We further design the training loss for mixed noise reduction. The experimental results show that the proposed LL-RefineNet outperforms the comparative methods both qualitatively and quantitatively with fast processing speed on both synthetic and natural low-light image datasets.


Author(s):  
Fernanda Melo Carneiro ◽  
João Paulo Francisco de Souza ◽  
Karina Dias Silva ◽  
Denis Silva Nogueira ◽  
David Bichsel ◽  
...  

The use of biodiversity surrogates is often suggested to increase the cost-effectiveness of biomonitoring programs, as this demands less time and taxonomic expertise. In addition, the detection of multi-taxon associations is a first step toward a better understanding of how organisms interact with each other. Such a multi-taxon association is termed a congruence, and can be detected through measuring the similarity in the distributional patterns shown by different biological groups. To assess the ability of different taxa to serve as surrogates for others, we carried out a Procrustes analysis on the beta diversity patterns of seven biological groups (aquatic birds, Amphibians, Macrophytes, Coleoptera, Odonata, Heteroptera and phytoplankton) in 35 ponds of the Cerrado biome. We found that: (i) the values of congruence in the studied ponds were weak; (ii) among all the biological groups compared, the highest congruence was found between amphibians and macrophytes; (iii) amphibians were congruent with the Coleoptera, Heteroptera, and macrophytes; (iv) the different taxa studied had different responses to environmental conditions; and (v) although they showed relatively weak congruence with the other taxa in each pond environment, amphibian communities were the most strongly influenced by environment variables. Almost all the communities observed in these systems showed unique pattern and thus should be studied and monitored in their entirety.


2013 ◽  
Vol 837 ◽  
pp. 727-732
Author(s):  
Nicoleta Acomi ◽  
Ovidiu Cristian Acomi ◽  
Alina Lucia Bostina ◽  
Aurel Bostina

The transport of aquatic organisms from one place to another by ships ballast water has created substantial environmental impact on discharging areas. In order to avoid that, the International Maritime Organization (IMO) recommends treating the ballast water by different methods. The paper presents four methods of managing ballast water, accepted by the Organization: treating by filtration and irradiation with ultraviolet light, treating by de-oxygenation, treating with biocide and treating by heating. The comparative analysis of the treating technologies allows the ship-owner to choose such equipment by analyzing the advantages and the disadvantages. Considering that not only the quality parameters are important but also the cost, the study is completed by a mathematical model for calculation of unit cost for treating ballast water onboard. The purpose of this study is to develop an instrument for selecting the optimum method of ballast water treatment suitable for each type of vessel or voyage, so the ship-owners would be able to choose a treatment method comparing the costs, based on the specific requirements.


2021 ◽  
Vol 11 (4) ◽  
pp. 3023-3029
Author(s):  
Muhammad Junaid ◽  
Luqman Shah ◽  
Ali Imran Jehangiri ◽  
Fahad Ali Khan ◽  
Yousaf Saeed ◽  
...  

With each passing day resolutions of still image/video cameras are on the rise. This amelioration in resolutions has the potential to extract useful information on the view opposite the photographed subjects from their reflecting parts. Especially important is the idea to capture images formed on the eyes of photographed people and animals. The motivation behind this research is to explore the forensic importance of the images/videos to especially analyze the reflections of the background of the camera. This analysis may include extraction/ detection/recognition of the objects in front of the subjects but on the back of the camera. In the national context such videos/photographs are not rare and, specifically speaking, an abductee’s video footage at a good resolution may give some important clues to the identity of the person who kidnapped him/her. Our aim would be to extract visual information formed in human eyes from still images as well as from video clips. After extraction, our next task would be to recognize the extracted visual information. Initially our experiments would be limited on characters’ extraction and recognition, including characters of different styles and font sizes (computerized) as well as hand written. Although varieties of Optical Character Recognition (OCR) tools are available for characters’ extraction and recognition but, the problem is that they only provide results for clear images (zoomed).


2001 ◽  
Vol 2001 (1) ◽  
pp. 213-217
Author(s):  
John C. Kern

ABSTRACT One challenge for trustees in a natural resource damage assessment (NRDA) is to adequately quantify natural resource injuries in a cost-effective manner. This is particularly true for smaller spills, where the cost of more expansive and more expensive injury assessment studies could dwarf the cost of the restoration actions to compensate for those injuries. The need for cost-effective assessments must he balanced against the need for the assessment methods to be technically defensible and useful in identifying and scaling appropriate restoration actions. In this paper, it is shown how the injury assessment results from the Lake Barre oil spill of May 1997 (which released 6,561 barrels of crude oil) were used to help inform trustees about the likely magnitude of injury for two smaller crude oil spills in Louisiana. For the Lake Barre spill, the trustees developed an incident-specific model—adapted from the Type A model—to quantify injury to birds and aquatic fauna. The results of this model were used to evaluate a restoration offer as compensation for these injuries from the responsible party (RP). Subsequently, the results of the Lake Barre assessment were used to help quantify injury to birds and aquatic organisms for the September 1998 release of up to 1,500 barrels of crude oil from a well blowout into Lake Grande Ecaille. The National Oceanic and Atmospheric Administration (NOAA) again used the Lake Barre results to quantify injury to water column organisms for a November 1999 release of 850 barrels of crude oil from a pipeline in Four-Bayou Pass. Estimating injury by extrapolation from one spill to similar spills represents one cost-effective approach toward quantifying injury for small incidents, and should be considered as a potential injury assessment method for those spills where it is impractical or otherwise difficult to justify conducting large incident-specific injury studies. This technique can be done quickly, potentially speeding the settlement and restoration implementation process, thereby compensating the public in an expeditious manner.


2020 ◽  
Vol 10 (8) ◽  
pp. 2794 ◽  
Author(s):  
Uduak Edet ◽  
Daniel Mann

A study to determine the visual requirements for a remote supervisor of an autonomous sprayer was conducted. Observation of a sprayer operator identified 9 distinct “look zones” that occupied his visual attention, with 39% of his time spent viewing the look zone ahead of the sprayer. While observation of the sprayer operator was being completed, additional GoPro cameras were used to record video of the sprayer in operation from 10 distinct perspectives (some look zones were visible from the operator’s seat, but other look zones were selected to display other regions of the sprayer that might be of interest to a sprayer operator). In a subsequent laboratory study, 29 experienced sprayer operators were recruited to view and comment on video clips selected from the video footage collected during the initial ride-along. Only the two views from the perspective of the operator’s seat were rated highly as providing important information even though participants were able to identify relevant information from all ten of the video clips. Generally, participants used the video clips to obtain information about the boom status, the location and movement of the sprayer within the field, the weather conditions (especially the wind), obstacles to be avoided, crop conditions, and field conditions. Sprayer operators with more than 15 years of experience provided more insightful descriptions of the video clips than their less experienced peers. Designers can influence which features the user will perceive by positioning the camera such that those specific features are prominent in the camera’s field of view. Overall, experienced sprayer operators preferred the concept of presenting visual information on an automation interface using live video rather than presenting that same information using some type of graphical display using icons or symbols.


2019 ◽  
Author(s):  
Yoram Harth

UNSTRUCTURED The high cost and scarcity of physicians results in lack of proper service to the majority of the population in the US and more so in the rest of the world. Telehealth, based on remote physicians does not seem to be the solution due to less than optimal cost/benefit ratio offered to the human provider. Recent developments in mobile processing power, mobile camera resolution, and deep learning technology present an opportunity to build solutions to specific diseases that are comparable in accuracy to a human in-person service for a fraction of the cost democratizing the availability of health services.


Sign in / Sign up

Export Citation Format

Share Document