document image binarization
Recently Published Documents


TOTAL DOCUMENTS

156
(FIVE YEARS 49)

H-INDEX

25
(FIVE YEARS 4)

2022 ◽  
Vol 70 (2) ◽  
pp. 3549-3564
Author(s):  
Wan Azani Mustafa ◽  
Haniza Yazid ◽  
Ahmed Alkhayyat ◽  
Mohd Aminudin Jamlos ◽  
Hasliza A. Rahim

2021 ◽  
Author(s):  
Carlos Garrido-Munoz ◽  
Adrián Sánchez-Hernández ◽  
Francisco J. Castellanos ◽  
Jorge Calvo-Zaragoza

Binarization represents a key role in many document image analysis workflows. The current state of the art considers the use of supervised learning, and specifically deep neural networks. However, it is very difficult for the same model to work successfully in a number of document styles, since the set of potential domains is very heterogeneous. We study a multi-source domain adaptation strategy for binarization. Within this scenario, we look into a novel hypothesis where a specialized binarization model must be selected to be used over a target domain, instead of a single model that tries to generalize across multiple domains. The problem then boils down to, given several specialized models and a new target set, deciding which model to use. We propose here a simple way to address this question by using a domain classifier, that estimates which of the source models must be considered to binarize the new target domain. Our experiments on several datasets, including different text styles and music scores, show that our initial hypothesis is quite promising, yet the way to deal with the decision of which model to use still shows great room for improvement.


2021 ◽  
pp. 1-14
Author(s):  
R.L. Jyothi ◽  
M. Abdul Rahiman

Binarization is the most important stage in historical document image processing. Efficient working of character and word recognition algorithms depend on effective segmentation methods. Segmentation algorithms in turn depend on images free of noises and degradations. Most of these historical documents are illegible with degradations like bleeding through degradation, faded ink or faint characters, uneven illumination, contrast variation, etc. For effective processing of these document images, efficient binarization algorithms should be devised. Here a simple modified version of the Convolutional Neural Network (CNN) is proposed for historical document binarization. AOD-Net architecture for generating dehazed images from hazed images is modified to create the proposed network.The new CNN model is created by incorporating Difference of Concatenation layer (DOC), Enhancement layer (EN) and Thresholding layer into AOD-Net to make it suitable for binarization of highly degraded document images. The DOC layer and EN layer work effectively in solving degradation that exists in the form of low pass noises. The complexity of working of the proposed model is reduced by decreasing the number of layers and by introducing filters in convolution layers that work with low inter-pixel dependency. This modified version of CNN works effectively with a variety of highly degraded documents when tested with the benchmark historical datasets. The main highlight of the proposed network is that it works efficiently in a generalized manner for any type of document images without further parameter tuning. Another important highlight of this method is that it can handle most of the degradation categories present in document images. In this work, the performance of the proposed model is compared with Otsu, Sauvola, and three recent Deep Learning-based models.


Author(s):  
Uche A. Nnolim

Conventional thresholding algorithms have had limited success with degraded document images. Recently, partial differential equations (PDEs) have been applied with good results. However, these are usually tailored to handle relatively few specific distortions. In this study, we combine an edge detection term with a linear binarization source term in a PDE formulation. Additionally, a new proposed diffusivity function further amplifies desired edges. It also suppresses undesired edges that comprise bleed-through effects. Furthermore, we develop the fractional variant of the proposed scheme, which further improves results and provides more flexibility. Moreover, nonlinear color spaces are utilized to improve binarization results for images with color distortion. The proposed scheme removes document image degradation such as bleed-through, stains, smudges, etc., and also restores faded text in the images. Experimental subjective and objective results show consistently superior performance of the proposed approach compared to the state-of-the-art PDE-based models.


Author(s):  
Mr. Aniket Pagare

Segmentation of text from badly degraded document images is an extremely difficult assignment because of the high inter/Intra variety between the record foundation and the frontal area text of various report pictures. Picture preparing and design acknowledgment algorithms set aside more effort for execution on a solitary center processor. Designs Preparing Unit (GPU) is more mainstream these days because of its speed, programmability, minimal expense and more inbuilt execution centers in it. The primary objective of this exploration work is to make binarization quicker for acknowledgment of a huge number of corrupted report pictures on GPU. In this framework, we give another picture division calculation that every pixel in the picture has its own limit proposed. We are accomplishing equal work on a window of m*n size and separate article pixel of text stroke of that window. The archive text is additionally sectioned by a nearby edge that is assessed dependent on the forces of identified content stroke edge pixels inside a nearby window.


2021 ◽  
pp. 108099
Author(s):  
Francisco J. Castellanos ◽  
Antonio-Javier Gallego ◽  
Jorge Calvo-Zaragoza

Sign in / Sign up

Export Citation Format

Share Document