optical character recognition
Recently Published Documents


TOTAL DOCUMENTS

1259
(FIVE YEARS 524)

H-INDEX

27
(FIVE YEARS 3)

Author(s):  
Karanrat Thammarak ◽  
Prateep Kongkla ◽  
Yaowarat Sirisathitkul ◽  
Sarun Intakosum

Optical character recognition (OCR) is a technology to digitize a paper-based document to digital form. This research studies the extraction of the characters from a Thai vehicle registration certificate via a Google Cloud Vision API and a Tesseract OCR. The recognition performance of both OCR APIs is also examined. The 84 color image files comprised three image sizes/resolutions and five image characteristics. For suitable image type comparison, the greyscale and binary image are converted from color images. Furthermore, the three pre-processing techniques, sharpening, contrast adjustment, and brightness adjustment, are also applied to enhance the quality of image before applying the two OCR APIs. The recognition performance was evaluated in terms of accuracy and readability. The results showed that the Google Cloud Vision API works well for the Thai vehicle registration certificate with an accuracy of 84.43%, whereas the Tesseract OCR showed an accuracy of 47.02%. The highest accuracy came from the color image with 1024×768 px, 300dpi, and using sharpening and brightness adjustment as pre-processing techniques. In terms of readability, the Google Cloud Vision API has more readability than the Tesseract. The proposed conditions facilitate the possibility of the implementation for Thai vehicle registration certificate recognition system.


Author(s):  
Rifiana Arief ◽  
Achmad Benny Mutiara ◽  
Tubagus Maulana Kusuma ◽  
Hustinawaty Hustinawaty

<p>This research proposed automated hierarchical classification of scanned documents with characteristics content that have unstructured text and special patterns (specific and short strings) using convolutional neural network (CNN) and regular expression method (REM). The research data using digital correspondence documents with format PDF images from pusat data teknologi dan informasi (technology and information data center). The document hierarchy covers type of letter, type of manuscript letter, origin of letter and subject of letter. The research method consists of preprocessing, classification, and storage to database. Preprocessing covers extraction using Tesseract optical character recognition (OCR) and formation of word document vector with Word2Vec. Hierarchical classification uses CNN to classify 5 types of letters and regular expression to classify 4 types of manuscript letter, 15 origins of letter and 25 subjects of letter. The classified documents are stored in the Hive database in Hadoop big data architecture. The amount of data used is 5200 documents, consisting of 4000 for training, 1000 for testing and 200 for classification prediction documents. The trial result of 200 new documents is 188 documents correctly classified and 12 documents incorrectly classified. The accuracy of automated hierarchical classification is 94%. Next, the search of classified scanned documents based on content can be developed.</p>


Author(s):  
Rajat Subhra Bhowmick ◽  
Isha Ganguli ◽  
Jayanta Paul ◽  
Jaya Sil

In today’s era of digitization, social media platforms play a significant role in networking and influencing the perception of the general population. Social network sites have recently been used to carry out harmful attacks against individuals, including political and theological figures, intellectuals, sports and movie stars, and other prominent dignitaries, which may or may not be intentional. However, the exchange of such information across the general population inevitably contributes to social-economic, socio-political turmoil, and even physical violence in society. By classifying the derogatory content of a social media post, this research work helps to eradicate and discourage the upsetting propagation of such hate campaigns. Social networking posts today often include the picture of Memes along with textual remarks and comments, which throw new challenges and opportunities to the research community while identifying the attacks. This article proposes a multimodal deep learning framework by utilizing ensembles of computer vision and natural language processing techniques to train an encapsulated transformer network for handling the classification problem. The proposed framework utilizes the fine-tuned state-of-the-art deep learning-based models (e.g., BERT, Electra) for multilingual text analysis along with face recognition and the optical character recognition model for Meme picture comprehension. For the study, a new Facebook meme-post dataset is created with recorded baseline results. The subject of the created dataset and context of the work is more geared toward multilingual Indian society. The findings demonstrate the efficacy of the proposed method in the identification of social media meme posts featuring derogatory content about a famous/recognized individual.


2022 ◽  
Vol 16 (1) ◽  
pp. 54
Author(s):  
Imam Husni Al amin ◽  
Awan Aprilino

Currently, vehicle number plate detection systems in general still use the manual method. This will take a lot of time and human effort. Thus, an automatic vehicle number plate detection system is needed because the number of vehicles that continues to increase will burden human labor. In addition, the methods used for vehicle number plate detection still have low accuracy because they depend on the characteristics of the object being used. This study develops a YOLO-based automatic vehicle number plate detection system. The dataset used is a pretrained YOLOv3 model of 700 data. Then proceed with the number plate text extraction process using the Tesseract Optical Character Recognition (OCR) library and the results obtained will be stored in the database. This system is web-based and API so that it can be used online and on the cross-platform. The test results show that the automatic number plate detection system reaches 100% accuracy with sufficient lighting and a threshold of 0.5 and for the results using the Tesseract library, the detection results are 92.32% where the system is successful in recognizing all characters on the license plates of cars and motorcycles. in the form of Alphanumeric characters of 7-8 characters.


2022 ◽  
Vol 20 (8) ◽  
pp. 3080
Author(s):  
A. A. Komkov ◽  
V. P. Mazaev ◽  
S. V. Ryazanova ◽  
D. N. Samochatov ◽  
E. V. Koshkina ◽  
...  

RuPatient health information system (HIS) is a computer program consisting of a doctor-patient web user interface, which includes algorithms for recognizing medical record text and entering it into the corresponding fields of the system.Aim. To evaluate the effectiveness of RuPatient HIS in actual clinical practice.Material and methods. The study involved 10 cardiologists and intensivists of the department of cardiology and сardiovascular intensive care unit of the L. A. Vorokhobov City Clinical Hospital 67 We analyzed images (scanned copies, photos) of discharge reports from patients admitted to the relevant departments in 2021. The following fields of medical documentation was recognized: Name, Complaints, Anamnesis of life and illness, Examination, Recommendations. The correctness and accuracy of recognition of entered information were analyzed. We compared the recognition quality of RuPatient HIS and a popular optical character recognition application (FineReader for Mac).Results. The study included 77 pages of discharge reports of patients from various hospitals in Russia from 50 patients (men, 52%). The mean age of patients was 57,7±7,9 years. The number of reports with correctly recognized fields in various categories using the program algorithms was distributed as follows: Name — 14 (28%), Diagnosis — 13 (26%), Complaints — 40 (80%), Anamnesis — 14 (28%), Examination — 24 (48%), Recommendations — 46 (92%). Data that was not included in the category was also recognized and entered in the comments field. The number of recognized words was 549±174,9 vs 522,4±215,6 (p=0,5), critical errors in words — 2,1±1,6 vs 4,4±2,8 (p<0,001), non-critical errors — 10,3±4,3 vs 5,6±3,3 (p<0,001) for RuPatient HIS and optical character recognition application for a personal computer, respectively.Conclusion. The developed RuPatient HIS, which includes a module for recognizing medical records and entering data into the corresponding fields, significantly increases the document management efficiency with high quality of optical character recognition based on neural network technologies and the automation of filling process.


Author(s):  
Armand Christopher Luna ◽  
Christian Trajano ◽  
John Paul So ◽  
Nicole John Pascua ◽  
Abraham Magpantay ◽  
...  

Author(s):  
Greta Franzini ◽  
Mike Kestemont ◽  
Gabriela Rotari ◽  
Melina Jander ◽  
Jeremi K. Ochab ◽  
...  

Az alábbi cikk egy multidiszciplináris projekt eredményeit mutatja be, amely a különböző digitalizációs stratégiák számítógépes szöveganalízisben való használhatóságát járja körül. Pontosabban Jacob és Wilhelm Grimm szerzőségének automatizált megkülönböztetésére tettünk kísérletet, melyet egy HTR (HandwrittenText Recognition – kézzel írott szöveg felismerése) és OCR (Optical Character Recognition – optikai karakterfelismerés) által feldolgozott levelezéskorpuszban hajtottunk végre, korrekció nélkül – felmérve, hogy az így keletkezett zaj milyen hatással van a fivérek különböző írásmódjának azonosítására. Összegezve,úgy tűnik, hogy az OCR megbízható helyettesítője lehet a manuális átírásnak, legalábbis a szerzőazonosítás kérdéskörét illetően. Eredményeink továbbá abba az irányba mutatnak, miszerint még a különböző digitalizációs eljárásokból származó tanító- és tesztkorpuszok (training and test set) is használhatók a szerzőazonosítás során. A HTR-t tekintve a kutatás azt demonstrálja, hogy ez az automatizált átírás ugyan az OCR-hez képest szignifikánsan növeli a szövegek félrecsoportosításának veszélyét, ám körülbelül 20% feletti tisztaság már önmagában elegendő ahhoz, hogy a véletlennél nagyobb esélye legyen a helyes binárismegfeleltetésnek.


METIK JURNAL ◽  
2021 ◽  
Vol 5 (2) ◽  
pp. 19-27
Author(s):  
Susana Lin ◽  
Genrawan Hoendarto

Financial management is one of the important things in the process of achieving the financial goals of a person or an organization. Everyone has their own way to manages finances, this is dependent on the character and they goals. Financial management can be done conventionally, for example by manual method which is commonly done by write the expenses, income, and savings in a notebook. However, if the note must contain details of the transactions carried out, it can be considered less efficient. The use of Optical Character Recognition will be able to answer this problem, by taking a picture of the transaction, all transaction details will be recorded on the smartphone, and the user can validate the results obtained and save the record on the smartphone user. Users can also immediately see the total transactions made according to the selected time range without having to calculate each transaction made manually. The application will be designed using the react native framework which allows it to run on various platforms.


2021 ◽  
Author(s):  
Tran Thi Anh Thu ◽  
Le Pham Ngoc Yen ◽  
Tran Thai Son ◽  
Dinh Dien

Sign in / Sign up

Export Citation Format

Share Document