sparql query
Recently Published Documents


TOTAL DOCUMENTS

208
(FIVE YEARS 67)

H-INDEX

16
(FIVE YEARS 3)

2021 ◽  
Vol 12 (1) ◽  
pp. 122
Author(s):  
Jongtae Lim ◽  
Byounghoon Kim ◽  
Hyeonbyeong Lee ◽  
Dojin Choi ◽  
Kyoungsoo Bok ◽  
...  

Various distributed processing schemes were studied to efficiently utilize a large scale of RDF graph in semantic web services. This paper proposes a new distributed SPARQL query processing scheme considering communication costs in Spark environments to reduce I/O costs during SPARQL query processing. We divide a SPARQL query into several subqueries using a WHERE clause to process a query of an RDF graph stored in a distributed environment. The proposed scheme reduces data communication costs by grouping the divided subqueries in related nodes through the index and processing them, and the grouped subqueries calculate the cost of all possible query execution paths to select an efficient query execution path. The efficient query execution path is selected through the algorithm considering the data parsing cost of all possible query execution paths, amount of data communication, and queue time per node. It is shown through various performance evaluations that the proposed scheme outperforms the existing schemes.


2021 ◽  
Author(s):  
Alberto Moya Loustaunau ◽  
Aidan Hogan
Keyword(s):  

Author(s):  
Usha Yadav ◽  
Neelam Duhan

With the evolution of Web 3.0, the traditional algorithm of searching Web 2.0 would become obsolete and underperform in retrieving the precise and accurate information from the growing semantic web. It is very reasonable to presume that common users might not possess any understanding of the ontology used in the knowledge base or SPARQL query. Therefore, providing easy access of this enormous knowledge base to all level of users is challenging. The ability for all level of users to effortlessly formulate structure query such as SPARQL is very diverse. In this paper, semantic web based search methodology is proposed which converts user query in natural language into SPARQL query, which could be directed to domain ontology based knowledge base. Each query word is further mapped to the relevant concept or relations in ontology. Score is assigned to each mapping to find out the best possible mapping for the query generation. Mapping with highest score are taken into consideration along with interrogative or other function to finally formulate the user query into SPARQL query. If there is no search result retrieved from the knowledge base, then instead of returning null to the user, the query is further directed to the Web 3.0. The top “k” documents are considered to further converting them into RDF format using Text2Onto tool and the corpus of semantically structured web documents is build. Alongside, semantic crawl agent is used to get <Subject-Predicate-Object> set from the semantic wiki. The Term Frequency Matrix and Co-occurrence Matrix are applied on the corpus following by singular Value decomposition (SVD) to find the results relevant for the user query. The result evaluations proved that the proposed system is efficient in terms of execution time, precision, recall and f-measures.


2021 ◽  
Author(s):  
Daniel Arturo Casal Amat ◽  
Carlos Buil-Aranda ◽  
Carlos Valle-Vidal

2021 ◽  
Author(s):  
Ayan Chatterjee

UNSTRUCTURED An automatic electronic coaching (eCoaching) can motivate individuals to lead a healthy lifestyle through early health risk prediction, customized recommendation generation, preference setting (such as, goal setting, response, and interaction), and goal evaluation. Such an eCoach system needs to collect heterogeneous health, wellness, and contextual data, and then convert them into meaningful information for health monitoring, health risk prediction, and the generation of personalized recommendations. However, data from various sources may cause a data compatibility dilemma. The proposed ontology can help in data integration, logical representation of sensory observations and customized suggestions, and discover implied knowledge. This "proof of concept (PoC)" research will help sensors, personal preferences, and recommendation data to be more organized. The research aims to design and develop an OWL-based ontology ("UiA Activity Recommendation Ontology/UiAARO") to annotate activity sensor data, contextual weather data, personal information, personal preferences, and personalized activity recommendations. The ontology was created using Protégé (V. 5.5.0) open-source software. We used the Java-based Jena Framework (V. 3.16) to build a semantic web application, which includes RDF API, OWL API, native tuple storage (TDB), and SPARQL query engine. The HermiT (V. 1.4.3.x) ontology reasoner available in Protégé 5.x has implemented the logical and structural consistency of the proposed ontology. The ontology can be visualized with OWLViz and OntoGraf, and the formal representation has been used to infer the health status of the eCoach participants with a reasoner. We have also developed an ontology verification module that behaves like a rule-based decision making (e.g., health state monitor and prediction), which can evaluate participant’s health state based on the evaluation of SPARQL query results, activity performed and predefined goal. Furthermore, the “UiAARO” has helped to represent the personalized recommendation messages beyond just “String” values, rather more meaningful with object-oriented representation. The scope of the proposed ontology is limited neither to specific sensor data nor only activity recommendations; instead, its scope can be further extended.


Semantic Web ◽  
2021 ◽  
pp. 1-19
Author(s):  
Marilena Daquino ◽  
Ivan Heibi ◽  
Silvio Peroni ◽  
David Shotton

Semantic Web technologies are widely used for storing RDF data and making them available on the Web through SPARQL endpoints, queryable using the SPARQL query language. While the use of SPARQL endpoints is strongly supported by Semantic Web experts, it hinders broader use of RDF data by common Web users, engineers and developers unfamiliar with Semantic Web technologies, who normally rely on Web RESTful APIs for querying Web-available data and creating applications over them. To solve this problem, we have developed RAMOSE, a generic tool developed in Python to create REST APIs over SPARQL endpoints. Through the creation of source-specific textual configuration files, RAMOSE enables the querying of SPARQL endpoints via simple Web RESTful API calls that return either JSON or CSV-formatted data, thus hiding all the intrinsic complexities of SPARQL and RDF from common Web users. We provide evidence that the use of RAMOSE to provide REST API access to RDF data within OpenCitations triplestores is beneficial in terms of the number of queries made by external users of such RDF data using the RAMOSE API, compared with the direct access via the SPARQL endpoint. Our findings show the importance for suppliers of RDF data of having an alternative API access service, which enables its use by those with no (or little) experience in Semantic Web technologies and the SPARQL query language. RAMOSE can be used both to query any SPARQL endpoint and to query any other Web API, and thus it represents an easy generic technical solution for service providers who wish to create an API service to access Linked Data stored as RDF in a triplestore.


2021 ◽  
Author(s):  
Sébastien Ferré

The results of a SPARQL query are generally presented as a table with one row per result, and one column per projected variable. This is an immediate consequence of the formal definition of SPARQL results as a sequence of mappings from variables to RDF terms. However, because of the flat structure of tables, some of the RDF graph structure is lost. This often leads to duplicates in the contents of the table, and difficulties to read and interpret results. We propose to use nested tables to improve the presentation of SPARQL results. A nested table is a table where cells may contain embedded tables instead of RDF terms, and so recursively. We introduce an automated procedure that lifts flat tables into nested tables, based on an analysis of the query. We have implemented the procedure on top of Sparklis, a guided query builder in natural language, in order to further improve the readability of its UI. It can as well be implemented on any SPARQL querying interface as it only depends on the query and its flat results. We illustrate our proposal in the domain of pharmacovigilance, and evaluate it on complex queries over Wikidata.


2021 ◽  
Author(s):  
Aisha Mohamed ◽  
Ghadeer Abuoda ◽  
Abdurrahman Ghanem ◽  
Zoi Kaoudi ◽  
Ashraf Aboulnaga

AbstractKnowledge graphs represented as RDF datasets are integral to many machine learning applications. RDF is supported by a rich ecosystem of data management systems and tools, most notably RDF database systems that provide a SPARQL query interface. Surprisingly, machine learning tools for knowledge graphs do not use SPARQL, despite the obvious advantages of using a database system. This is due to the mismatch between SPARQL and machine learning tools in terms of data model and programming style. Machine learning tools work on data in tabular format and process it using an imperative programming style, while SPARQL is declarative and has as its basic operation matching graph patterns to RDF triples. We posit that a good interface to knowledge graphs from a machine learning software stack should use an imperative, navigational programming paradigm based on graph traversal rather than the SPARQL query paradigm based on graph patterns. In this paper, we present RDFFrames, a framework that provides such an interface. RDFFrames provides an imperative Python API that gets internally translated to SPARQL, and it is integrated with the PyData machine learning software stack. RDFFrames enables the user to make a sequence of Python calls to define the data to be extracted from a knowledge graph stored in an RDF database system, and it translates these calls into a compact SPQARL query, executes it on the database system, and returns the results in a standard tabular format. Thus, RDFFrames is a useful tool for data preparation that combines the usability of PyData with the flexibility and performance of RDF database systems.


2021 ◽  
Vol 11 (15) ◽  
pp. 7033
Author(s):  
Oscar Ceballos ◽  
Carlos Alberto Ramírez Restrepo ◽  
María Constanza Pabón ◽  
Andres M. Castillo ◽  
Oscar Corcho

Existing SPARQL query engines and triple stores are continuously improved to handle more massive datasets. Several approaches have been developed in this context proposing the storage and querying of RDF data in a distributed fashion, mainly using the MapReduce Programming Model and Hadoop-based ecosystems. New trends in Big Data technologies have also emerged (e.g., Apache Spark, Apache Flink); they use distributed in-memory processing and promise to deliver higher data processing performance. In this paper, we present a formal interpretation of some PACT transformations implemented in the Apache Flink DataSet API. We use this formalization to provide a mapping to translate a SPARQL query to a Flink program. The mapping was implemented in a prototype used to determine the correctness and performance of the solution. The source code of the project is available in Github under the MIT license.


Sign in / Sign up

Export Citation Format

Share Document