Scene text reading based cloud compliance access

2020 ◽  
Vol 23 (4) ◽  
pp. 2633-2647
Author(s):  
Hezhong Pan ◽  
Chuanyi Liu ◽  
Shaoming Duan ◽  
Peiyi Han ◽  
Binxing Fang
Sensors ◽  
2021 ◽  
Vol 21 (5) ◽  
pp. 1919
Author(s):  
Shuhua Liu ◽  
Huixin Xu ◽  
Qi Li ◽  
Fei Zhang ◽  
Kun Hou

With the aim to solve issues of robot object recognition in complex scenes, this paper proposes an object recognition method based on scene text reading. The proposed method simulates human-like behavior and accurately identifies objects with texts through careful reading. First, deep learning models with high accuracy are adopted to detect and recognize text in multi-view. Second, datasets including 102,000 Chinese and English scene text images and their inverse are generated. The F-measure of text detection is improved by 0.4% and the recognition accuracy is improved by 1.26% because the model is trained by these two datasets. Finally, a robot object recognition method is proposed based on the scene text reading. The robot detects and recognizes texts in the image and then stores the recognition results in a text file. When the user gives the robot a fetching instruction, the robot searches for corresponding keywords from the text files and achieves the confidence of multiple objects in the scene image. Then, the object with the maximum confidence is selected as the target. The results show that the robot can accurately distinguish objects with arbitrary shape and category, and it can effectively solve the problem of object recognition in home environments.


2014 ◽  
Vol 36 (2) ◽  
pp. 375-387 ◽  
Author(s):  
Jerod J. Weinman ◽  
Zachary Butler ◽  
Dugan Knoll ◽  
Jacqueline Feild
Keyword(s):  

2019 ◽  
Vol 87 ◽  
pp. 118-129 ◽  
Author(s):  
Dinh NguyenVan ◽  
Shijian Lu ◽  
Shangxuan Tian ◽  
Nizar Ouarti ◽  
Mounir Mokhtari
Keyword(s):  

2020 ◽  
Vol 10 (13) ◽  
pp. 4474 ◽  
Author(s):  
Direselign Addis Tadesse ◽  
Chuan-Ming Liu ◽  
Van-Dai Ta

Reading text and unified text detection and recognition from natural images are the most challenging applications in computer vision and document analysis. Previously proposed end-to-end scene text reading methods do not consider the frequency of input images at feature extraction, which slows down the system, requires more memory, and recognizes text inaccurately. In this paper, we proposed an octave convolution (OctConv) feature extractor and a time-restricted attention encoder-decoder module for end-to-end scene text reading. The OctConv can extract features by factorizing the input image based on their frequency. It is a direct replacement of convolutions, orthogonal and complementary, for reducing redundancies and helps to boost the reading text through low memory requirements at a faster speed. In the text reading process, features are first extracted from the input image using Feature Pyramid Network (FPN) with OctConv Residual Network with depth 50 (ResNet50). Then, a Region Proposal Network (RPN) is applied to predict the location of the text area by using extracted features. Finally, a time-restricted attention encoder-decoder module is applied after the Region of Interest (RoI) pooling is performed. A bilingual real and synthetic scene text dataset is prepared for training and testing the proposed model. Additionally, well-known datasets including ICDAR2013, ICDAR2015, and Total Text are used for fine-tuning and evaluating its performance with previously proposed state-of-the-art methods. The proposed model shows promising results on both regular and irregular or curved text detection and reading tasks.


2020 ◽  
Vol 34 (07) ◽  
pp. 11899-11907 ◽  
Author(s):  
Liang Qiao ◽  
Sanli Tang ◽  
Zhanzhan Cheng ◽  
Yunlu Xu ◽  
Yi Niu ◽  
...  

Many approaches have recently been proposed to detect irregular scene text and achieved promising results. However, their localization results may not well satisfy the following text recognition part mainly because of two reasons: 1) recognizing arbitrary shaped text is still a challenging task, and 2) prevalent non-trainable pipeline strategies between text detection and text recognition will lead to suboptimal performances. To handle this incompatibility problem, in this paper we propose an end-to-end trainable text spotting approach named Text Perceptron. Concretely, Text Perceptron first employs an efficient segmentation-based text detector that learns the latent text reading order and boundary information. Then a novel Shape Transform Module (abbr. STM) is designed to transform the detected feature regions into regular morphologies without extra parameters. It unites text detection and the following recognition part into a whole framework, and helps the whole network achieve global optimization. Experiments show that our method achieves competitive performance on two standard text benchmarks, i.e., ICDAR 2013 and ICDAR 2015, and also obviously outperforms existing methods on irregular text benchmarks SCUT-CTW1500 and Total-Text.


2012 ◽  
Vol 71 (3) ◽  
pp. 141-148 ◽  
Author(s):  
Doriane Gras ◽  
Hubert Tardieu ◽  
Serge Nicolas

Predictive inferences are anticipations of what could happen next in the text we are reading. These inferences seem to be activated during reading, but a delay is necessary for their construction. To determine the length of this delay, we first used a classical word-naming task. In the second experiment, we used a Stroop-like task to verify that inference activation was not due to strategies applied during the naming task. The results show that predictive inferences are naturally activated during text reading, after approximately 1 s.


Author(s):  
Tobias Alf Kroll ◽  
A. Alexandre Trindade ◽  
Amber Asikis ◽  
Melissa Salas ◽  
Marcy Lau ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document