Linking Page Images to Transcriptions with SVG

Author(s):  
Hugh A. Cayless

This paper will present the results of ongoing experimentation with the linking of manuscript images to TEI transcriptions. The method being tested involves the automated conversion of images containing text to SVG, using Open Source tools. Once the text has been converted to SVG paths, these can be grouped in the document to mark the words therein and these groups can then be linked using standard methods to tokenized versions of the transcriptions. The goal of these experiments is to achieve a much more fine-grained linking and annotation mechanism than is so far possible with available tools, e.g. the Image Markup Tool and TEI P5 facsimile markup, both of which annotate only rectangular sections of an image. The method envisioned here would produce a legible tracing of the word, expressed in XML, to which transcripts and annotations might be attached and which can be superimposed upon the original image.

2011 ◽  
Vol 403-408 ◽  
pp. 1933-1936
Author(s):  
Ke Qiang Ren ◽  
Hai Ying Jiang

JPEG2000 is a new compressed standard of static image, Kakadu is a open source system which can be higher efficiency to realize the algorithm of JPEG2000. This article introduces the codec construction and the compressed stream structure in JPEG2000, analyzes the construction of Kakadu and image compression class in Kakadu, and then carries on application experiments based on JPEG2000 in Kakadu 2.2 platform. Experiments show that the compression of interest image region has higher transfer rate and lower memory compared with the original image, and JPEG2000 has higher compression rate and better visual quality compared with JPEG.


2021 ◽  
Vol 2142 (1) ◽  
pp. 012013
Author(s):  
A S Nazdryukhin ◽  
A M Fedrak ◽  
N A Radeev

Abstract This work presents the results of using self-normalizing neural networks with automatic selection of hyperparameters, TabNet and NODE to solve the problem of tabular data classification. The method of automatic selection of hyperparameters was realised. Testing was carried out with the open source framework OpenML AutoML Benchmark. As part of the work, a comparative analysis was carried out with seven classification methods, experiments were carried out for 39 datasets with 5 methods. NODE shows the best results among the following methods and overperformed standard methods for four datasets.


Author(s):  
Amanjot Singh ◽  
Jagroop Singh

Image Super resolution is used to upscale the low resolution Images. It is also known as image upscaling .This paper focuses on upscaling of compressed images based on Interpolation techniques. A content adaptive interpolation method of image upscaling has been proposed. This interpolation based scheme is useful for single image based Super-resolution (SR) methods .The presented method works on horizontal, vertical and diagonal directions of an image separately and it is adaptive to the local content of an image. This method relies only on single image and uses the content of the original image only; therefore the proposed method is more practical and realistic. The simulation results have been compared to other standard methods with the help of various performance matrices like PSNR, MSE, MSSIM etc. which indicates the preeminence of the proposed method.


Database ◽  
2019 ◽  
Vol 2019 ◽  
Author(s):  
Guilherme Viteri ◽  
Lisa Matthews ◽  
Thawfeek Varusai ◽  
Marc Gillespie ◽  
Marija Milacic ◽  
...  

Abstract Reactome is a manually curated, open-source, open-data knowledge base of biomolecular pathways. Reactome has always provided clear credit attribution for authors, curators and reviewers through fine-grained annotation of all three roles at the reaction and pathway level. These data are visible in the web interface and provided through the various data download formats. To enhance visibility and credit attribution for the work of authors, curators and reviewers, and to provide additional opportunities for Reactome community engagement, we have implemented key changes to Reactome: contributor names are now fully searchable in the web interface, and contributors can ‘claim’ their contributions to their ORCID profile with a few clicks. In addition, we are reaching out to domain experts to request their help in reviewing and editing Reactome pathways through a new ‘Contribution’ section, highlighting pathways which are awaiting community review. Database URL: https://reactome.org


2020 ◽  
Author(s):  
Christine Blume ◽  
Christian Cajochen

The detection of sleep cycles in human sleep data (i.e. polysomnographically assessed sleep stages) enables fine-grained analyses of ultradian variations in sleep microstructure (e.g. sleep spindles, and arousals), or other amplitude- and frequency-specific electroencephalographic features during sleep. While many laboratories have software that is used internally, reproducibility requires the availability of open source software. Therefore, we here introduce the ‘SleepCycles’ package for R, an open-source software package that identifies sleep cycles and their respective (non-) rapid eye movement ([N]REM) periods from sleep staging data. Additionally, each (N)REM period is subdivided into parts of equal duration, which may be useful for further fine-grained analyses. The detection criteria are, with some adaptations, largely based on criteria originally proposed by Feinberg and Floyd (1979). The latest version of the package can be downloaded from the Comprehensive R Archives Network (CRAN).•The package ‘SleepCycles’ for R allows to identify sleep cycles and their respective NREM and REM from sleep staging results.•Besides the cycle detection, NREM and REM are also split into parts of equal duration (percentiles) thereby allowing for a better temporal resolution across the night and temporal alignment of sleep cycles with different durations among different night recordings.


Cybersecurity ◽  
2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Gu Ban ◽  
Lili Xu ◽  
Yang Xiao ◽  
Xinhua Li ◽  
Zimu Yuan ◽  
...  

AbstractCodes of Open Source Software (OSS) are widely reused during software development nowadays. However, reusing some specific versions of OSS introduces 1-day vulnerabilities of which details are publicly available, which may be exploited and lead to serious security issues. Existing state-of-the-art OSS reuse detection work can not identify the specific versions of reused OSS well. The features they selected are not distinguishable enough for version detection and the matching scores are only based on similarity.This paper presents B2SMatcher, a fine-grained version identification tool for OSS in commercial off-the-shelf (COTS) software. We first discuss five kinds of version-sensitive code features that are trackable in both binary and source code. We categorize these features into program-level features and function-level features and propose a two-stage version identification approach based on the two levels of code features. B2SMatcher also identifies different types of OSS version reuse based on matching scores and matched feature instances. In order to extract source code features as accurately as possible, B2SMatcher innovatively uses machine learning methods to obtain the source files involved in the compilation and uses function abstraction and normalization methods to eliminate the comparison costs on redundant functions across versions. We have evaluated B2SMatcher using 6351 candidate OSS versions and 585 binaries. The result shows that B2SMatcher achieves a high precision up to 89.2% and outperforms state-of-the-art tools. Finally, we show how B2SMatcher can be used to evaluate real-world software and find some security risks in practice.


Author(s):  
Venkatesh Naidu Nerella ◽  
Martin Krause ◽  
Viktor Mechtcherine

Buildability, i.e. the ability of a deposited material bulk to retain its dimmensions under increasing load, is an inherent prerequisite for formwork-free digital construction (DC). Since DC processes are relatively new, no standard methods of characterization are available yet. The paper at hand presents practice-oriented buildabilty criteria by taking various process parameters and construction costs into consideration. In doing so, direct links between laboratory buildability tests and target applications are established. A systematic basis for calculating the time interval (TI) to be followed during laboratory testing is proposed for the full-width printing (FWP) and filament printing (FP) processes. The proposed approach is validated by applying it to a high-strength, printable, fine-grained concrete. Comparative analyses of FWP and FP revealed that to test the buildability of a material for FP processes, higher velocities of the printhead should be established for laboratory tests in comparison to those needed for FWP process, providing for equal construction rates.


2020 ◽  
Vol 11 (1) ◽  
pp. 103-115
Author(s):  
Grant Snitker

Sedimentary charcoal analysis is increasingly used in archaeological and paleoecological research to examine human-environmental relationships at multiple scales. The recent availability of low-cost digital microscopes and imaging software has resulted in the rapid adoption of digital image analysis in charcoal studies. However, most published studies include only minimal accounts of software configurations or utilize proprietary image analysis programs, thus hindering replication, standardization, and comparability of charcoal analyses across the field. In an effort to encourage replicable methods and a culture of open science, this paper presents the Charcoal Quantification Tool (CharTool), a free, open-source suite of charcoal and sediment quantification tools designed for use with ImageJ. CharTool blends standard methods in visual and digital charcoal analysis to increase the analyst’s participation in identifying and measuring charcoal metrics. Each CharTool module is described and demonstrated in a vignette using sedimentary charcoal collected from the Son Servera study area, Mallorca, Spain. A suggested workflow, user-guide, scripted analyses for processing outputs, and download instructions are included as supplementary materials to this article.


Author(s):  
N. Börlin ◽  
A. Murtiyoso ◽  
P. Grussenmeyer

Abstract. The Damped Bundle Adjustment Toolbox (DBAT) is a free, open-source, toolbox for bundle adjustment. The purpose of DBAT is to provide an independent, open-source toolkit for statistically rigorous bundle adjustment computations. The capabilities include bundle adjustment, network analysis, point filtering, forward intersection, spatial intersection, plotting functions, and computations of quality indicators such as posterior covariance estimates and parameter correlations. DBAT is written in the high-level Matlab language and includes several processing example files. The input formats have so far been restricted to PhotoModeler export files and Photoscan (Metashape) native files. Fine-tuning of the processing has so far required knowledge of the Matlab language.This paper describes the development of a scripting language based on the XML (eXtensible Markup Language) language that allow the user a fine-grained control over what operations are applied to the input data, while keeping the needed programming skills at a minimum. Furthermore, the scripting language allows a wide range of input formats. Additionally, the XML format allows simple extension of the script file format both in terms of adding new operations, file formats, or adding parameters to existing operations. Overall, the script files will in principle allow DBAT to process any kind of photogrammetric input and should extend the usability of DBAT as a scientific and teaching tool for photogrammetric computations.


Sign in / Sign up

Export Citation Format

Share Document