scholarly journals Handwritten Script Recognition using DCT, Gabor Filter and Wavelet Features at Line Level

Author(s):  
Dr. G.G. Rajput ◽  
Anita H.B.

In a country like India where more number of scripts are in use, automatic identification of printed and handwritten script facilitates many important applications including sorting of document images and searching online archives of document images. In this paper, a multiple feature based approach is presented to identify the script type of the collection of handwritten documents. Eight popular Indian scripts are considered here. Features are extracted using Gabor filters, Discrete Cosine Transform, and Wavelets of Daubechies family. Experiments are performed to test the recognition accuracy of the proposed system at line level for bilingual scripts and later extended to trilingual scripts. We have obtained 100% recognition accuracy for bi-scripts at line level. The classification is done using k-nearest neighbour classifier.

2019 ◽  
Vol 39 (5) ◽  
pp. 0528004
Author(s):  
李非燕 Li Feiyan ◽  
霍宏涛 Huo Hongtao ◽  
李静 Li Jing ◽  
白杰 Bai Jie

2014 ◽  
Vol 23 (3) ◽  
pp. 245-260 ◽  
Author(s):  
Ram Sarkar ◽  
Nibaran Das ◽  
Subhadip Basu ◽  
Mahantapas Kundu ◽  
Mita Nasipuri

AbstractA novel piecewise water flow technique for text line extraction from multi-skewed document images of handwritten text of different scripts is presented here. The basic water flow technique assumes that the hypothetical water flows from both left and right sides of the image frame. This flow of water fills up the gaps between consecutive objects (texts) but faces obstruction if any object lies in the path of the flow. All unwetted regions in the document image are then labeled distinctly to extract the text lines. However, the technique fails when two neighboring text lines touch each other, as water gets obstructed by the touching segment(s). To get rid of this difficulty, we have modified the basic water flow technique by iteratively applying the same over the vertically segmented document images. The main purpose of this vertical segmentation is to localize the text line segment(s) where two text lines get joined. These segments are then horizontally fragmented, and each fragment is placed suitably to the text line in which it actually belongs to. This way, the probable data loss during isolation of the touching text line segment is minimized. Both the techniques (current and basic ones) have been tested on three different databases, viz., CMATERdb 1.1.1, CMATERdb 1.1.2, and ICDAR2009 handwritten segmentation contest pages, respectively. The test results show that the present technique outperforms the basic one for all three databases.


Author(s):  
Johannes Erfurt ◽  
Wang-Q Lim ◽  
Heiko Schwarz ◽  
Detlev Marpe ◽  
Thomas Wiegand

Abstract Recent progress in video compression is seemingly reaching its limits making it very hard to improve coding efficiency significantly further. The adaptive loop filter (ALF) has been a topic of interest for many years. ALF reaches high coding gains and has motivated many researchers over the past years to further improve the state-of-the-art algorithms. The main idea of ALF is to apply a classification to partition the set of all sample locations into multiple classes. After that, Wiener filters are calculated and applied for each class. Therefore, the performance of ALF essentially relies on how its classification behaves. In this paper, we extensively analyze multiple feature-based classifications for ALF (MCALF) and extend the original MCALF by incorporating sample adaptive offset filtering. Furthermore, we derive new block-based classifications which can be applied in MCALF to reduce its complexity. Experimental results show that our extended MCALF can further improve compression efficiency compared to the original MCALF algorithm.


2011 ◽  
Vol 20 (03) ◽  
pp. 489-509 ◽  
Author(s):  
BEHZAD HELLI ◽  
MOHSEN EBRAHIMI MOGHADDAM

The behavioral-biometrics methods of writer identification and verification have been considered as a research topic for many years. However, many writer identification and verification methods have been designed based on English handwriting properties, but because of many differences between English and Persian handwriting and the challenges facing Persian handwriting analysis, designing such methods has many interests in Persian yet. In this paper, we have presented a fully text-independent and texture based method for identifying writers of Persian handwritten documents. As a result of special properties of Persian handwriting, a modified version of Gabor filter that is called Extended Gabor (XGabor) filter has been used to extract the features. An MLP (Multi Layer Perceptron (Node)) neural network and a K-NN classifier have been employed to classify the extracted features. In the evaluation phase, an exhaustive database of Persian handwritten documents was prepared and the method applied on. The experimental results showed that the accuracy of proposed method is about 97% and it is competitive with others. We believe that the proposed method may be extended to identify writers in other languages by adjusting some parameters.


2020 ◽  
Author(s):  
Leandro Tacioli ◽  
Luíz Toledo ◽  
Claudua Medeiros

Automatic identification of animals is extremely useful for scientists, providing ways to monitor species and changes in ecological communities. The choice of effective audio features and classification techniques is a challenge on any audio recognition system, especially in bioacoustics that commonly uses several algorithms. This paper presents a novel software architecture that supports multiple feature extraction and classification algorithms to help on the identification of animal species from their recorded sounds. This architecture was implemented by the WASIS software, freely available on the Web.


Historical documents are important source for knowing culture, language, social activities, educational system, etc. The historical documents are in different languages and evolved over centuries and transformed to present modern language, classification of documents into various eras, recognition of words etc. In this paper, we have proposed a new approach to automatic identification of the age of the historical handwritten document images based on LBP (Local Binary Pattern) and LPQ (Local Phase Quantization) algorithm. The standard historical handwritten document images named as MPS (Medieval Paleographic Scale) dataset which is publicly available is used to experiment. LBP and LPQ descriptors are used to extract the features of the historical document images. Further, documents are classified based on the discriminating feature values using classifiers namely K-NN (K-Nearest Neighbors) and SVM (Support Vector Machine) classifier. The accuracy of historical handwritten document images by K-NN and SVM are 90.7% and 92.8% respectively.


Sensors ◽  
2020 ◽  
Vol 20 (10) ◽  
pp. 2914
Author(s):  
Hubert Michalak ◽  
Krzysztof Okarma

Image binarization is one of the key operations decreasing the amount of information used in further analysis of image data, significantly influencing the final results. Although in some applications, where well illuminated images may be easily captured, ensuring a high contrast, even a simple global thresholding may be sufficient, there are some more challenging solutions, e.g., based on the analysis of natural images or assuming the presence of some quality degradations, such as in historical document images. Considering the variety of image binarization methods, as well as their different applications and types of images, one cannot expect a single universal thresholding method that would be the best solution for all images. Nevertheless, since one of the most common operations preceded by the binarization is the Optical Character Recognition (OCR), which may also be applied for non-uniformly illuminated images captured by camera sensors mounted in mobile phones, the development of even better binarization methods in view of the maximization of the OCR accuracy is still expected. Therefore, in this paper, the idea of the use of robust combined measures is presented, making it possible to bring together the advantages of various methods, including some recently proposed approaches based on entropy filtering and a multi-layered stack of regions. The experimental results, obtained for a dataset of 176 non-uniformly illuminated document images, referred to as the WEZUT OCR Dataset, confirm the validity and usefulness of the proposed approach, leading to a significant increase of the recognition accuracy.


Sign in / Sign up

Export Citation Format

Share Document