line segmentation
Recently Published Documents


TOTAL DOCUMENTS

265
(FIVE YEARS 81)

H-INDEX

21
(FIVE YEARS 2)

Author(s):  
Xinglong Wang ◽  
Jinde Zheng ◽  
Jun Zhang

Abstract The level selection of frequency band division structure relies on previous information in most gram approaches that capture the optimal demodulation frequency band (ODFB). When an improper level is specified in these approaches, the fault characteristic information contained in the produced ODFB may be insufficient. This research proposes a unique approach termed median line-gram (MELgram) to tackle the level selection problem. To divide the frequency domain signal into a series of demodulation frequency bands, a spectrum median line segmentation model based on Akima interpolation is first created. The level and boundary of the segmentation model can be adaptively identified by this means. Second, the acquired frequency bands are quantized using the negative entropy index, and the ODFB is defined as the frequency band with the largest value. Third, the envelope spectrum is used to determine the ODFB characteristic frequency to pinpoint the bearing fault location. Finally, both simulation and experimental signal analysis are used to demonstrate the efficiency of the suggested method. Furthermore, the suggested method extracts more defect feature information from the ODFB than existing methods.


2021 ◽  
Author(s):  
Xiuguang Song ◽  
Zhaoyou Ma ◽  
Chunyu Zhou ◽  
Fei Cheng ◽  
Xucai Zhuang ◽  
...  

2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Benjamin Kiessling ◽  
Gennady Kurin ◽  
Matthew Thomas Miller ◽  
Kader Smail

This work presents an accuracy study of the open source OCR engine, Kraken, on the leading Arabic scholarly journal, al-Abhath. In contrast with other commercially available OCR engines, Kraken is shown to be capable of producing highly accurate Arabic-script OCR. The study also assesses the relative accuracy of typeface-specific and generalized models on the al-Abhath data and provides a microanalysis of the “error instances” and the contextual features that may have contributed to OCR misrecognition. Building on this analysis, the paper argues that Arabic-script OCR can be significantly improved through (1) a more systematic approach to training data production, and (2) the development of key technological components, especially multi-language models and improved line segmentation and layout analysis./Cet article présente une étude d’exactitude du moteur ROC open source, Krakan, sur la revue académique arabe de premier rang, al-Abhath. Contrairement à d’autres moteurs ROC disponibles sur le marché, Kraken se révèle être capable de produire de la ROC extrêmement exacte de l’écriture arabe. L’étude évalue aussi l’exactitude relative des modèles spécifiquement configurés à des polices et celle des modèles généralisés sur les données d’al-Abhath et fournit une microanalyse des « occurrences d’erreurs », ainsi qu’une microanalyse des éléments contextuels qui pourraient avoir contribué à la méreconnaissance ROC. S’appuyant sur cette analyse, cet article fait valoir que la ROC de l’écriture arabe peut être considérablement améliorée grâce à (1) une approche plus systématique d’entraînement de la production de données et (2) grâce au développement de composants technologiques fondamentaux, notammentl’amélioration des modèles multilingues, de la segmentation de ligne et de l’analyse de la mise en page.


Author(s):  
Manoj Kumar Dixit

Text detection in video frames provide highly condensed information about the content of the video and it is useful for video seeking, browsing, retrieval and understanding video text in large video databases. In this paper, we propose a hybrid method that it automatically detects segments and recognizes the text present in the video. Detection is done by using laplacian method based on wavelet and color features. Segmentation of detected text is divided into two modules Line segmentation and Character segmentation. Line segmentation is done by using mathematical statistical method based on projection profile analysis. In line segmentation, multiple lines of text in video frame obtained from text detection are segmented into single line. Character segmentation is done by using Connected Component. Analysis (CCA) and Vertical Projection Profile Analysis. The input for character segmentation is the line of text obtained from line segmentation, in which all the characters in the line are segmented separately for recognition. Optical character recognition is Processed by using template matching and correlation technique. Template matching is performed by comparing an input character with a set of templates, each comparison results in a similarity measure between the input characters with a set of templates. After all templates have been compared with the observed character image, the character’s identity is assigned with the most similar template based on correlation. Eventually, the text in video frame is detected, segmented, and processed to OCR for recognition.


Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Kang Liu ◽  
Xin Gao

The use of multimodal sensors for lane line segmentation has become a growing trend. To achieve robust multimodal fusion, we introduced a new multimodal fusion method and proved its effectiveness in an improved fusion network. Specifically, a multiscale fusion module is proposed to extract effective features from data of different modalities, and a channel attention module is used to adaptively calculate the contribution of the fused feature channels. We verified the effect of multimodal fusion on the KITTI benchmark dataset and A2D2 dataset and proved the effectiveness of the proposed method on the enhanced KITTI dataset. Our method achieves robust lane line segmentation, which is 4.53% higher than the direct fusion on the precision index, and obtains the highest F2 score of 79.72%. We believe that our method introduces an optimization idea of modal data structure level for multimodal fusion.


2021 ◽  
Vol 2021 (HistoInformatics) ◽  
Author(s):  
Pit Schneider

Text line segmentation is one of the pre-stages of modern optical character recognition systems. The algorithmic approach proposed by this paper has been designed for this exact purpose. Its main characteristic is the combination of two different techniques, morphological image operations and horizontal histogram projections. The method was developed to be applied on a historic data collection that commonly features quality issues, such as degraded paper, blurred text, or presence of noise. For that reason, the segmenter in question could be of particular interest for cultural institutions, that want access to robust line bounding boxes for a given historic document. Because of the promising segmentation results that are joined by low computational cost, the algorithm was incorporated into the OCR pipeline of the National Library of Luxembourg, in the context of the initiative of reprocessing their historic newspaper collection. The general contribution of this paper is to outline the approach and to evaluate the gains in terms of accuracy and speed, comparing it to the segmentation algorithm bundled with the used open source OCR software.


Author(s):  
Zhenhong Zou ◽  
Xinyu Zhang ◽  
Huaping Liu ◽  
Zhiwei Li ◽  
Amir Hussain ◽  
...  

Author(s):  
Frederic Chaume

Whereas new technologies seem to favour globalisation in many areas of translation, dubbing shows a reluctance to embrace this trend of globalisation. Translation memories are now used to make translation easier and faster all over the world. In the area of audiovisual translation, new subtitling software has been developed, which is now widely used among both practitioners and companies. Also in subtitling, most microtextual practices (line segmentation, subtitle segmentation, typographical usages, synthesis of information, etc.) are followed by the majority of professionals. But dubbing seems to refuse to bend to homogenisation. Perhaps due to notions of nationalism and singularity attached to this concept, dubbing still shows different macro- and microtextual practices in the European countries in which it is the most popular type of audiovisual translation.This paper examines four different dubbing practices at a microtextual level -those carried out in Germany, Italy, Spain and France. Before considering new failed attempts to globalise this practice, and also some major advances brought by new technologies, the paper focuses on the differences in layout, take segmentation and dialogue writing in these four countries. These differences show that dubbing practices are still very conservative, and resistant to change and homogenisation.


Sign in / Sign up

Export Citation Format

Share Document