Automated Recognition of Numeric Display Based on Deep Learning

Author(s):  
Bohdan Rusyn ◽  
Oleksiy Lutsyk ◽  
Rostyslav Kosarevych ◽  
Yarema Varetsky
2021 ◽  
Author(s):  
Hayley Weir ◽  
Keiran Thompson ◽  
Amelia Woodward ◽  
Benjamin Choi ◽  
Augustin Braun ◽  
...  

Inputting molecules into chemistry software, such as quantum chemistry packages, currently requires domain expertise, expensive software and/or cumbersome procedures. Leveraging recent breakthroughs in machine learning, we develop ChemPix: an offline,...


Author(s):  
Kulendu Kashyap Chakraborty ◽  
Rashmi Mukherjee ◽  
Chandan Chakroborty ◽  
Kangkana Bora

2020 ◽  
Vol 10 (3) ◽  
pp. 359-367 ◽  
Author(s):  
Amin Khouani ◽  
Mostafa El Habib Daho ◽  
Sidi Ahmed Mahmoudi ◽  
Mohammed Amine Chikh ◽  
Brahim Benzineb

Diversity ◽  
2020 ◽  
Vol 12 (1) ◽  
pp. 29 ◽  
Author(s):  
Alina Raphael ◽  
Zvy Dubinsky ◽  
David Iluz ◽  
Nathan S. Netanyahu

We present thorough this review the developments in the field, point out their current limitations, and outline its timelines and unique potential. In order to do so we introduce the methods used in each of the advances in the application of deep learning (DL) to coral research that took place between the years: 2016–2018. DL has unique capability of streamlining the description, analysis, and monitoring of coral reefs, saving time, and obtaining higher reliability and accuracy compared with error-prone human performance. Coral reefs are the most diverse and complex of marine ecosystems, undergoing a severe decline worldwide resulting from the adverse synergistic influences of global climate change, ocean acidification, and seawater warming, exacerbated by anthropogenic eutrophication and pollution. DL is an extension of some of the concepts originating from machine learning that join several multilayered neural networks. Machine learning refers to algorithms that automatically detect patterns in data. In the case of corals these data are underwater photographic images. Based on “learned” patterns, such programs can recognize new images. The novelty of DL is in the use of state-of-art computerized image analyses technologies, and its fully automated methodology of dealing with large data sets of images. Automated Image recognition refers to technologies that identify and detect objects or attributes in a digital video or image automatically. Image recognition classifies data into selected categories out of many. We show that Neural Network methods are already reliable in distinguishing corals from other benthos and non-coral organisms. Automated recognition of live coral cover is a powerful indicator of reef response to slow and transient changes in the environment. Improving automated recognition of coral species, DL methods already recognize decline of coral diversity due to natural and anthropogenic stressors. Diversity indicators can document the effectiveness of reef bioremediation initiatives. We explored the current applications of deep learning for corals and benthic image classification by discussing the most recent studies conducted by researchers. We review the developments in the field, point out their current limitations, and outline their timelines and unique potential. We also discussed a few future research directions in the fields of deep learning. Future needs are the age detection of single species, in order to track trends in their population recruitment, decline, and recovery. Fine resolution, at the polyp level, is still to be developed, in order to allow separation of species with similar macroscopic features. That refinement of DL will allow such comparisons and their analyses. We conclude that the usefulness of future, more refined automatic identification will allow reef comparison, and tracking long term changes in species diversity. The hitherto unused addition of intraspecific coral color parameters, will add the inclusion of physiological coral responses to environmental conditions and change thereof. The core aim of this review was to underscore the strength and reliability of the DL approach for documenting coral reef features based on an evaluation of the currently available published uses of this method. We expect that this review will encourage researchers from computer vision and marine societies to collaborate on similar long-term joint ventures.


2020 ◽  
Author(s):  
Yanhua Gao ◽  
Yuan Zhu ◽  
Bo Liu ◽  
Yue Hu ◽  
Youmin Guo

ObjectiveIn Transthoracic echocardiographic (TTE) examination, it is essential to identify the cardiac views accurately. Computer-aided recognition is expected to improve the accuracy of the TTE examination.MethodsThis paper proposes a new method for automatic recognition of cardiac views based on deep learning, including three strategies. First, A spatial transform network is performed to learn cardiac shape changes during the cardiac cycle, which reduces intra-class variability. Second, a channel attention mechanism is introduced to adaptively recalibrates channel-wise feature responses. Finally, unlike conventional deep learning methods, which learned each input images individually, the structured signals are applied by a graph of similarities among images. These signals are transformed into the graph-based image embedding, which act as unsupervised regularization constraints to improve the generalization accuracy.ResultsThe proposed method was trained and tested in 171792 cardiac images from 584 subjects. Compared with the known result of the state of the art, the overall accuracy of the proposed method on cardiac image classification is 99.10% vs. 91.7%, and the mean AUC is 99.36%. Moreover, the overall accuracy is 98.15%, and the mean AUC is 98.96% on an independent test set with 34211 images from 100 subjects.ConclusionThe method of this paper achieved the results of the state of the art, which is expected to be an automated recognition tool for cardiac views recognition. The work confirms the potential of deep learning on ultrasound medicine.


2020 ◽  
Vol 5 (2) ◽  
pp. 24-34 ◽  
Author(s):  
Tino Mager ◽  
Carola Hein

Digital technologies provide novel ways of visualizing cities and buildings. They also facilitate new methods of analyzing the built environment, ranging from artificial intelligence (AI) to crowdsourced citizen participation. Digital representations of cities have become so refined that they challenge our perception of the real. However, computers have not yet become able to detect and analyze the visible features of built structures depicted in photographs or other media. Recent scientific advances mean that it is possible for this new field of computer vision to serve as a critical aid to research. Neural networks now meet the challenge of identifying and analyzing building elements, buildings and urban landscapes. The development and refinement of these technologies requires more attention, simultaneously, investigation is needed in regard to the use and meaning of these methods for historical research. For example, the use of AI raises questions about the ways in which computer-based image recognition reproduces biases of contemporary practice. It also invites reflection on how mixed methods, integrating quantitative and qualitative approaches, can be established and used in research in the humanities. Finally, it opens new perspectives on the role of crowdsourcing in both knowledge dissemination and shared research. Attempts to analyze historical big data with the latest methods of deep learning, to involve many people—laymen and experts—in research via crowdsourcing and to deal with partly unknown visual material have provided a better understanding of what is possible. The article presents findings from the ongoing research project ArchiMediaL, which is at the forefront of the analysis of historical mediatizations of the built environment. It demonstrates how the combination of crowdsourcing, historical big data and deep learning simultaneously raises questions and provides solutions in the field of architectural and urban planning history.


Author(s):  
C. Pezzica ◽  
J. Schroeter ◽  
O. E. Prizeman ◽  
C. B. Jones ◽  
P. L. Rosin

<p><strong>Abstract.</strong> Building on the richness of recent contributions in the field, this paper presents a state-of-the-art CNN analysis method for automating the recognition of standardised building components in modern heritage buildings. At the turn of the twentieth century manufactured building components became widely advertised for specification by architects. Consequently, a form of standardisation across various typologies began to take place. During this era of rapid economic and industrialised growth, many forms of public building were erected. This paper seeks to demonstrate a method for informing the recognition of such elements using deep learning to recognise ‘families’ of elements across a range of buildings in order to retrieve and recognise their technical specifications from the contemporary trade literature. The method is illustrated through the case of Carnegie Public Libraries in the UK, which provides a unique but ubiquitous platform from which to explore the potential for the automated recognition of manufactured standard architectural components. The aim of enhancing this knowledge base is to use the degree to which these were standardised originally as a means to inform and so support their ongoing care but also that of many other contemporary buildings. Although these libraries are numerous, they are maintained at a local level and as such, their shared challenges for maintenance remain unknown to one another. Additionally, this paper presents a methodology to indirectly retrieve useful indicators and semantics, relating to emerging HBIM families, by applying deep learning to a varied range of architectural imagery.</p>


2021 ◽  
Author(s):  
Kohulan Rajan ◽  
Henning Otto brinkhaus ◽  
Maria Sorokina ◽  
Achim Zielesny ◽  
Christoph Steinbeck

<p>Chemistry looks back at many decades of publications on chemical compounds, their structures and properties, in scientific articles. Liberating this knowledge (semi-)automatically and making it available to the world in open-access databases is a current challenge. Apart from mining textual information, Optical Chemical Structure Recognition (OCSR), the translation of an image of a chemical structure into a machine-readable representation, is part of this workflow. As the OCSR process requires an image containing a chemical structure, there is a need for a publicly available tool that automatically recognizes and segments chemical structure depictions from scientific publications. This is especially important for older documents which are only available as scanned pages. Here, we present DECIMER (Deep lEarning for Chemical IMagE Recognition) Segmentation, the first open-source, deep learning-based tool for automated recognition and segmentation of chemical structures from the scientific literature.</p><br><p>The workflow is divided into two main stages. During the detection step, a deep learning model recognizes chemical structure depictions and creates masks which define their positions on the input page. Subsequently, potentially incomplete masks are expanded in a post-processing workflow. The performance of DECIMER Segmentation has been manually evaluated on three sets of publications from different publishers. The approach operates on bitmap images of journal pages to be applicable also to older articles before the introduction of vector images in PDFs. </p><br><p>By making the source code and the trained model publicly available, we hope to contribute to the development of comprehensive chemical data extraction workflows. In order to facilitate access to DECIMER Segmentation, we also developed a web application. The web application, available at <a href="https://decimer.ai">https://decimer.ai</a>, lets the user upload a pdf file and retrieve the segmented structure depictions.</p><div><br></div>


2021 ◽  
Vol 21 (1) ◽  
Author(s):  
Wen Pan ◽  
Xujia Li ◽  
Weijia Wang ◽  
Linjing Zhou ◽  
Jiali Wu ◽  
...  

Abstract Background Development of a deep learning method to identify Barrett's esophagus (BE) scopes in endoscopic images. Methods 443 endoscopic images from 187 patients of BE were included in this study. The gastroesophageal junction (GEJ) and squamous-columnar junction (SCJ) of BE were manually annotated in endoscopic images by experts. Fully convolutional neural networks (FCN) were developed to automatically identify the BE scopes in endoscopic images. The networks were trained and evaluated in two separate image sets. The performance of segmentation was evaluated by intersection over union (IOU). Results The deep learning method was proved to be satisfying in the automated identification of BE in endoscopic images. The values of the IOU were 0.56 (GEJ) and 0.82 (SCJ), respectively. Conclusions Deep learning algorithm is promising with accuracies of concordance with manual human assessment in segmentation of the BE scope in endoscopic images. This automated recognition method helps clinicians to locate and recognize the scopes of BE in endoscopic examinations.


Sign in / Sign up

Export Citation Format

Share Document