Semantic Extension of Query for the Linked Data

2017 ◽  
Vol 13 (4) ◽  
pp. 109-133 ◽  
Author(s):  
Pu Li ◽  
Yuncheng Jiang ◽  
Ju Wang ◽  
Zhilei Yin

With the advent of Big Data Era, users prefer to get knowledge rather than pages from Web. Linked Data, a new form of knowledge representation and publishing described by RDF, can provide a more precise and comprehensible semantic structure to satisfy the aforementioned requirement. Further, the SPARQL query language for RDF is the foundation of many current researches about Linked Data querying. However, these SPARQL-based methods cannot fully express the semantics of the query, so they cannot unleash the potential of Linked Data. To fill this gap, this paper designs a new querying method which extends the SPARQL pattern. Firstly, the authors present some new semantic properties for predicates in RDF triples and design a Semantic Matrix for Predicates (SMP). They then establish a well-defined framework for the notion of Semantically-Extended Query Model for the Linked Data (SEQMLD). Moreover, the authors propose some novel algorithms for executing queries by integrating semantic extension into SPARQL pattern. Lastly, experimental results show that the authors' proposal has a good generality and performs better than some of the most representative similarity search methods.

Author(s):  
Martins Zviedris

End-user interaction with data is one of key aspects in data processing. Nowadays a lot of information systems have a custom made user interface for data input and data querying. From 1970s it is envisioned that a generic, user-friendly approach for data querying could be built, but no wide spread solution has been developed. In the paper we present a diagrammatic query language. We have done an iterative approach to design and improve the diagrammatic query language to make it user readable. Readability is analyzed with questionnaires. Readable diagrammatic query language is the first step to create a more generic and user-friendly data querying.


2009 ◽  
pp. 961-986
Author(s):  
Franck Ravat ◽  
Olivier Teste ◽  
Gilles Zurfluh

This chapter deals with constraint-based multidimensional modelling. The model we define integrates a constellation of facts and dimensions. Along each dimension, various hierarchies are possibly defined and the model supports multiple instantiations of dimensions. The main contribution is the definition of intra-dimension constraints between hierarchies of a same dimension as well as inter-dimension constraints of various dimensions. To facilitate data querying, we define a multi-dimensional query algebra, which integrates the main multi-dimensional operators such as rotations, drill down, roll up… These operators support the constraint-based multi-dimensional modelling. Finally, we present two implementations of this algebra. First, OLAP-SQL is a textual language integrating multi-dimensional concepts (fact, dimension, hierarchy), but it is based on classical SQL syntax. This language is dedicated to specialists such as multi-dimensional database administrators. Second, a graphical query language is presented. This language consists in a graphical representation of multi-dimensional databases, and users specify directly their queries over this graph. This approach is dedicated to non-computer scientist users.


Author(s):  
Franck Ravat ◽  
Olivier Teste ◽  
Gilles Zurfluh

This chapter deals with constraint-based multi-dimensional modelling. The model we define integrates a constellation of facts and dimensions. Along each dimension, various hierarchies are possibly defined and the model supports multiple instantiations of dimensions. The main contribution is the definition of intra-dimension constraints between hierarchies of a same dimension as well as inter-dimension constraints of various dimensions. To facilitate data querying, we define a multi-dimensional query algebra, which integrates the main multi-dimensional operators such as rotations, drill down, roll up… These operators support the constraint-based multi-dimensional modelling. Finally, we present two implementations of this algebra. First, OLAP-SQL is a textual language integrating multi-dimensional concepts (fact, dimension, hierarchy), but it is based on classical SQL syntax. This language is dedicated tospecialists such as multi-dimensional database administrators. Second, a graphical query language is presented. This language consists in a graphical representation of multi-dimensional databases, and users specify directly their queries over this graph. This approach is dedicated to non-computer scientist users.


2003 ◽  
Vol 60 ◽  
Author(s):  
Roberta Pires De Oliveira

Várias propriedades sintáticas e semânticas de todo N, chamadas bare universal phrase (BUP), e todos(s) o(s) N(s), defined universal phrase (DUP), podem ser explicadas atribuindo-se uma estrutura semântica a cada sintagma quantificacional. Se nossa proposta semântica estiver correta, torna-se incorreta a afirmação de Peres (1992) de que a presença do artigo definido no sintagma quantificacional é nula. As estruturas semânticas atribuídas a BUP e DUP justificam a hipótese de que todo N não é um quantificador. Tal hipótese encontra ressonância na proposição de Negrão (2002), que afirma ser todo N um indefinido, no sentido de Heim. As análises de Dayal (1998) e Saeboe (2001) acerca de any são usadas para mostrar que todo N é um universal, não um indedinido. Além disso, trata-se de um modal. A discussão a respeito da generalização de Enç é feita para demonstrar que, uma vez que se considere todo como quantificador como a melhor solução, tal generalização deverá ser revista. Nossa proposta é a de manter a distinção entre a distinção entre quantificação e especificidade. A especificidade é dada pela presença do artigo definido. O quantificador propriamente dito expressa quantificação. Tal proposta abre a possibilidade de uma classe de quantificadores não-específicos (talvez modais), dos quais todo N é um exemplo central. Abstract Several syntactic and semantic properties of todo N, named bare universal phrase (BUP), and todo(s) o(s) N(s), defined universal phrase (DUP), may be explained by attributing to each quantifier phrase a semantic structure. If our semantic proposal is correct, then Peres’s (1992) claim that the presence of the definite article in the quantifier phrase is vacuous is not right. The semantic structures attributed to BUP and DUP lead to the hypothesis that todo N was not a quantifier. Such a hypothesis finds resonance in Negrão’s proporal (2002) that claims that todo N is an indefinite in Heim’s sense. Dayal’s (1998) and Saeboe’s (2001) analyses of any are used to show that todo N is a universal, not an indefinite. Moreover it is a modal. The discussion of Enç’s generalization is brought up in order to show that if considering todo as a quantifier is a better solution, then Enç’s generalization needs to be revised. Our proposal is to keep apart quantification from specificity. Specificity is given by the presence of the definite article. The quantifier itself expresses quantification. Such a proposal opens up the possibility of a class of non-specific (perhaps modal) quantifiers, of which todo N is a central example.


2018 ◽  
Vol 10 (8) ◽  
pp. 2613
Author(s):  
Dandan He ◽  
Zhongfu Li ◽  
Chunlin Wu ◽  
Xin Ning

Industrialized construction has raised the requirements of procurement methods used in the construction industry. The rapid development of e-commerce offers efficient and effective solutions, however the large number of participants in the construction industry means that the data involved are complex, and problems arise related to volume, heterogeneity, and fragmentation. Thus, the sector lags behind others in the adoption of e-commerce. In particular, data integration has become a barrier preventing further development. Traditional e-commerce platform, which considered data integration for common product data, cannot meet the requirements of construction product data integration. This study aimed to build an information-integrated e-commerce platform for industrialized construction procurement (ICP) to overcome some of the shortcomings existing platforms. We proposed a platform based on Building Information Modelling (BIM) and linked data, taking an innovative approach to data integration. It uses industrialized construction technology to support product standardization, BIM to support procurement process, and linked data to connect different data sources. The platform was validated using a case study. With the development of an e-commerce ontology, industrialized construction component information was extracted from BIM models and converted to Resource Description Framework (RDF) format. Related information from different data sources was also converted to RDF format, and Simple Protocol and Resource Description Framework Query Language (SPARQL) queries were implemented. The platform provides a solution for the development of e-commerce platform in the construction industry.


2015 ◽  
Vol 31 (3) ◽  
pp. 415-429 ◽  
Author(s):  
Loredana Di Consiglio ◽  
Tiziana Tuoto

Abstract The Capture-recapture method is a well-known solution for evaluating the unknown size of a population. Administrative data represent sources of independent counts of a population and can be jointly exploited for applying the capture-recapture method. Of course, administrative sources are affected by over- or undercoverage when considered separately. The standard Petersen approach is based on strong assumptions, including perfect record linkage between lists. In reality, record linkage results can be affected by errors. A simple method for achieving linkage error-unbiased population total estimates is proposed in Ding and Fienberg (1994). In this article, an extension of the Ding and Fienberg model by relaxing their conditions is proposed. The procedures are illustrated for estimating the total number of road casualties, on the basis of a probabilistic record linkage between two administrative data sources. Moreover, a simulation study is developed, providing evidence that the adjusted estimator always performs better than the Petersen estimator.


2017 ◽  
Author(s):  
Alexander Garcia ◽  
Federico Lopez ◽  
Leyla Garcia ◽  
Olga Giraldo ◽  
Victor Bucheli ◽  
...  

A significant portion of biomedical literature is represented in a manner that makes it difficult for consumers to find or aggregate content through a computational query. One approach to facilitate reuse of the scientific literature is to structure this information as linked data using standardized web technologies. In this paper we present the second version of Biotea, a semantic, linked data version of the open-access subset of PubMed Central that has been enhanced with specialized annotation pipelines that uses existing infrastructure from the National Center for Biomedical Ontology. We expose our models, services, software and datasets. Our infrastructure enables manual and semi-automatic annotation, resulting data are represented as RDF-based linked data and can be readily queried using the SPARQL query language. We illustrate the utility of our system with several use cases. Availability: Our datasets, methods and techniques are available at http://biotea.github.io


2016 ◽  
Vol 54 (11) ◽  
pp. 2726-2734 ◽  
Author(s):  
Claudia R. Molins ◽  
Mark J. Delorey ◽  
Christopher Sexton ◽  
Martin E. Schriefer

The current recommendation for the laboratory confirmation of Lyme disease is serology-based diagnostics. Specifically, a standardized two-tiered testing (STTT) algorithm is applied that utilizes a first-tier immunofluorescence assay or enzyme immunoassay (EIA) that, if the result is positive or equivocal, is followed by second-tier immunoblotting. Despite the standardization and performance achievements, STTT is considered technically complex and subjective, as well as insensitive for early acute infection. These issues have prompted development of novel algorithms and testing platforms. In this study, we evaluated the performance of several commonly used assays for STTT. Several modified two-tiered testing (MTTT) algorithms, including a 2-EIA algorithm and modified criteria for second-tier IgG immunoblots, were also evaluated. All tests were performed on sera from a recently available, well-defined archive of positive- and negative-control patients. Our study demonstrates differences in the results between individual first- and second-tier tests, although the overall agreement of the different STTT algorithms used was strong. In addition, the MTTT algorithm utilizing 2-EIAs was found to be equivalent to all STTT algorithms tested, with agreement ranging from 94 to 97%. The 2-EIA MTTT algorithm slightly enhanced sensitivity in early disease compared to the STTT algorithms evaluated. Furthermore, these data add to the mounting evidence that a 2-EIA-based MTTT algorithm, where immunoblotting is replaced by the C6 EIA, performs as well or better than STTT.


Sign in / Sign up

Export Citation Format

Share Document