CCANet: Exploiting Pixel-wise Semantics for Irregular Scene Text Spotting

Author(s):  
Shanbo Xu ◽  
Chen Chen ◽  
Silong Peng ◽  
Xiyuan Hu
Keyword(s):  
2021 ◽  
Author(s):  
Yizhang Huang ◽  
Kun Fang ◽  
Xiaolin Huang ◽  
Jie Yang

Author(s):  
Wei Feng ◽  
Fei Yin ◽  
Xu-Yao Zhang ◽  
Wenhao He ◽  
Cheng-Lin Liu

2020 ◽  
Vol 34 (07) ◽  
pp. 12160-12167 ◽  
Author(s):  
Hao Wang ◽  
Pu Lu ◽  
Hui Zhang ◽  
Mingkun Yang ◽  
Xiang Bai ◽  
...  

Recently, end-to-end text spotting that aims to detect and recognize text from cluttered images simultaneously has received particularly growing interest in computer vision. Different from the existing approaches that formulate text detection as bounding box extraction or instance segmentation, we localize a set of points on the boundary of each text instance. With the representation of such boundary points, we establish a simple yet effective scheme for end-to-end text spotting, which can read the text of arbitrary shapes. Experiments on three challenging datasets, including ICDAR2015, TotalText and COCO-Text demonstrate that the proposed method consistently surpasses the state-of-the-art in both scene text detection and end-to-end text recognition tasks.


2021 ◽  
Author(s):  
Yu Zhou ◽  
Hongtao Xie ◽  
Shancheng Fang ◽  
Jing Wang ◽  
Zhengjun Zha ◽  
...  

2021 ◽  
pp. 1-11
Author(s):  
Guangcun Wei ◽  
Wansheng Rong ◽  
Yongquan Liang ◽  
Xinguang Xiao ◽  
Xiang Liu

Aiming at the problem that the traditional OCR processing method ignores the inherent connection between the text detection task and the text recognition task, This paper propose a novel end-to-end text spotting framework. The framework includes three parts: shared convolutional feature network, text detector and text recognizer. By sharing convolutional feature network, the text detection network and the text recognition network can be jointly optimized at the same time. On the one hand, it can reduce the computational burden; on the other hand, it can effectively use the inherent connection between text detection and text recognition. This model add the TCM (Text Context Module) on the basis of Mask RCNN, which can effectively solve the negative sample problem in text detection tasks. This paper propose a text recognition model based on the SAM-BiLSTM (spatial attention mechanism with BiLSTM), which can more effectively extract the semantic information between characters. This model significantly surpasses state-of-the-art methods on a number of text detection and text spotting benchmarks, including ICDAR 2015, Total-Text.


2021 ◽  
pp. 288-299
Author(s):  
Beiji Zou ◽  
Wenjun Yang ◽  
Kaiwen Li ◽  
Enquan Huang ◽  
Shu Liu

Sign in / Sign up

Export Citation Format

Share Document