Large-Scale Optical Character Recognition of Ancient Greek

2017 ◽  
Vol 14 (3) ◽  
pp. 341-359 ◽  
Author(s):  
Bruce Robertson ◽  
Federico Boschetti
2012 ◽  
Vol 6 (1-2) ◽  
pp. 111-119 ◽  
Author(s):  
Elspeth Haston ◽  
Robert Cubey ◽  
David J. Harris

Logistically, the data associated with biological collections can be divided into three main categories for digitisation: i) Label Data: the data appearing on the specimen on a label or annotation; ii) Curatorial Data: the data appearing on containers, boxes, cabinets and folders which hold the collections; iii) Supplementary Data: the data held separately from the collections in indices, archives and literature. Each of these categories of data have fundamentally different properties within the digitisation framework which have implications for the data capture process. These properties were assessed in relation to alternative data entry workflows and methodologies to create a more efficient and accurate system of data capture. We see a clear benefit in the prioritisation of curatorial data in the data capture process. These data are often only available at the cabinets, they are in a format suitable for allowing rapid data entry, and they result in an accurate cataloguing of the collections. Finally, the capture of a high resolution digital image enables additional data entry to be separated into multiple sweeps, and optical character recognition (OCR) software can be used to facilitate sorting images for fuller data entry, and giving potential for more automated data entry in the future.


2014 ◽  
Vol 42 (2) ◽  
pp. 336-350 ◽  
Author(s):  
Jim Hahn

Purpose – The purpose of this paper is to report results of a formative usability study that investigated first-year student use of an optical character recognition (OCR) mobile application (app) designed to help students find resources for course assignments. The app uses textual content from the assignment sheet to suggest relevant library resources of which students may not be aware. Design/methodology/approach – Formative evaluation data are collected to inform the production level version of the mobile application and to understand student use models and requirements for OCR software in mobile applications. Findings – Mobile OCR apps are helpful for undergraduate students searching known titles of books, general subject areas or searching for help guide content developed by the library. The results section details how student feedback shaped the next iteration of the app for integration as a Minrva module. Research limitations/implications – This usability paper is not a large-scale quantitative study, but seeks to provide deep qualitative research data for the specific mobile interface studied, the Text-shot prototype. Practical implications – The OCR application is designed to help students learn about availability of library resources based on scanning (e.g. taking a picture, or “Text-shot”) of an assignment sheet, a course syllabus or other course-related handouts. Originality/value – This study contributes a new area of application development for libraries, with research methods that are useful for other mobile development studies.


1997 ◽  
Vol 9 (1-3) ◽  
pp. 58-77
Author(s):  
Vitaly Kliatskine ◽  
Eugene Shchepin ◽  
Gunnar Thorvaldsen ◽  
Konstantin Zingerman ◽  
Valery Lazarev

In principle, printed source material should be made machine-readable with systems for Optical Character Recognition, rather than being typed once more. Offthe-shelf commercial OCR programs tend, however, to be inadequate for lists with a complex layout. The tax assessment lists that assess most nineteenth century farms in Norway, constitute one example among a series of valuable sources which can only be interpreted successfully with specially designed OCR software. This paper considers the problems involved in the recognition of material with a complex table structure, outlining a new algorithmic model based on ‘linked hierarchies’. Within the scope of this model, a variety of tables and layouts can be described and recognized. The ‘linked hierarchies’ model has been implemented in the ‘CRIPT’ OCR software system, which successfully reads tables with a complex structure from several different historical sources.


2020 ◽  
Vol 2020 (1) ◽  
pp. 78-81
Author(s):  
Simone Zini ◽  
Simone Bianco ◽  
Raimondo Schettini

Rain removal from pictures taken under bad weather conditions is a challenging task that aims to improve the overall quality and visibility of a scene. The enhanced images usually constitute the input for subsequent Computer Vision tasks such as detection and classification. In this paper, we present a Convolutional Neural Network, based on the Pix2Pix model, for rain streaks removal from images, with specific interest in evaluating the results of the processing operation with respect to the Optical Character Recognition (OCR) task. In particular, we present a way to generate a rainy version of the Street View Text Dataset (R-SVTD) for "text detection and recognition" evaluation in bad weather conditions. Experimental results on this dataset show that our model is able to outperform the state of the art in terms of two commonly used image quality metrics, and that it is capable to improve the performances of an OCR model to detect and recognise text in the wild.


Sign in / Sign up

Export Citation Format

Share Document