Schema-level Index Models for Web Data Search

2021 ◽  
Vol 2 (1) ◽  
pp. 47-63
Author(s):  
Ansgar Scherp ◽  
Till Blume

Indexing the Web of Data offers many opportunities, in particular, to find and explore data sources. One major design decision when indexing the Web of Data is to find a suitable index model, i.e., how to index and summarize data. Various efforts have been conducted to develop specific index models for a given task. With each index model designed, implemented, and evaluated independently, it remains difficult to judge whether an approach generalizes well to another task, set of queries, or dataset. In this work, we empirically evaluate six representative index models with unique feature combinations. Among them is a new index model incorporating inferencing over RDFS and \texttt{owl:sameAs}. We implement all index models for the first time into a single, stream-based framework. We evaluate variations of the index models considering sub-graphs of size $0$, $1$, and $2$ hops on two large, real-world datasets. We evaluate the quality of the indices regarding the compression ratio, summarization ratio, and F1-score denoting the approximation quality of the stream-based index computation. The experiments reveal huge variations in compression ratio, summarization ratio, and approximation quality for different index models, queries, and datasets. However, we observe meaningful correlations in the results that help to determine the right index model for a given task, type of query, and dataset.


Author(s):  
Muhammad Ahtisham Aslam ◽  
Naif Radi Aljohani

Producing the Linked Open Data (LOD) is getting potential to publish high-quality interlinked data. Publishing such data facilitates intelligent searching from the Web of data. In the context of scientific publications, data about millions of scientific documents published by hundreds and thousands of publishers is in silence as it is not published as open data and ultimately is not linked to other datasets. In this paper the authors present SPedia: a semantically enriched knowledge base of data about scientific documents. SPedia knowledge base provides information on more than nine million scientific documents, consisting of more than three hundred million RDF triples. These extracted datasets, allow users to put sophisticated queries by employing semantic Web techniques instead of relying on keyword-based searches. This paper also shows the quality of extracted data by performing sample queries through SPedia SPARQL Endpoint and analyzing results. Finally, the authors describe that how SPedia can serve as central hub for the cloud of LOD of scientific publications.



2017 ◽  
Vol 13 (1) ◽  
pp. 128-147 ◽  
Author(s):  
Muhammad Ahtisham Aslam ◽  
Naif Radi Aljohani

Producing the Linked Open Data (LOD) is getting potential to publish high-quality interlinked data. Publishing such data facilitates intelligent searching from the Web of data. In the context of scientific publications, data about millions of scientific documents published by hundreds and thousands of publishers is in silence as it is not published as open data and ultimately is not linked to other datasets. In this paper the authors present SPedia: a semantically enriched knowledge base of data about scientific documents. SPedia knowledge base provides information on more than nine million scientific documents, consisting of more than three hundred million RDF triples. These extracted datasets, allow users to put sophisticated queries by employing semantic Web techniques instead of relying on keyword-based searches. This paper also shows the quality of extracted data by performing sample queries through SPedia SPARQL Endpoint and analyzing results. Finally, the authors describe that how SPedia can serve as central hub for the cloud of LOD of scientific publications.



2019 ◽  
Author(s):  
FRANCISCO CARLOS PALETTA

This work aims to presents partial results on the research project conducted at the Observatory of the Labor Market in Information and Documentation, School of Communications and Arts of the University of São Paulo on Information Science and Digital Humanities. Discusses Digital Humanities and informational literacy. Highlights the evolution of the Web, the digital library and its connections with Digital Humanities. Reflects on the challenges of the Digital Humanities transdisciplinarity and its connections with the Information Science. This is an exploratory study, mainly due to the current and emergence of the theme and the incipient bibliography existing both in Brazil and abroad.Keywords: Digital Humanities; Information Science; Transcisciplinrity; Information Literacy; Web of Data; Digital Age.



2019 ◽  
Vol 54 (6) ◽  
Author(s):  
Sawsan Ali Hamid ◽  
Rana Alauldeen Abdalrahman ◽  
Inam Abdullah Lafta ◽  
Israa Al Barazanchi

Recently, web services have presented a new and evolving model for constructing the distributed system. The meteoric growth of the Web over the last few years proves the efficacy of using simple protocols over the Internet as the basis for a large number of web services and applications. Web service is a modern technology of web, which can be defined as software applications with a programmatic interface based on Internet protocol. Web services became common in the applications of the web by the help of Universal, Description, Discovery and Integration; Web Service Description Language and Simple Object Access Protocol. The architecture of web services refers to a collection of conceptual components in which common sets of standard can be defined among interoperating components. Nevertheless, the existing Web service's architecture is not impervious to some challenges, such as security problems, and the quality of services. Against this backdrop, the present study will provide an overview of these issues. Therefore, it aims to propose web services architecture model to support distributed system in terms of application and issues.





Author(s):  
Axel Polleres ◽  
David Huynh Huynh


2017 ◽  
Author(s):  
Stephen Richard ◽  
◽  
Douglas Fils ◽  
Anders Noren ◽  
Kerstin A. Lehnert
Keyword(s):  


Sensors ◽  
2021 ◽  
Vol 21 (5) ◽  
pp. 1817
Author(s):  
Jiawen Xue ◽  
Li Yin ◽  
Zehua Lan ◽  
Mingzhu Long ◽  
Guolin Li ◽  
...  

This paper proposes a novel 3D discrete cosine transform (DCT) based image compression method for medical endoscopic applications. Due to the high correlation among color components of wireless capsule endoscopy (WCE) images, the original 2D Bayer data pattern is reconstructed into a new 3D data pattern, and 3D DCT is adopted to compress the 3D data for high compression ratio and high quality. For the low computational complexity of 3D-DCT, an optimized 4-point DCT butterfly structure without multiplication operation is proposed. Due to the unique characteristics of the 3D data pattern, the quantization and zigzag scan are ameliorated. To further improve the visual quality of decompressed images, a frequency-domain filter is proposed to eliminate the blocking artifacts adaptively. Experiments show that our method attains an average compression ratio (CR) of 22.94:1 with the peak signal to noise ratio (PSNR) of 40.73 dB, which outperforms state-of-the-art methods.



2021 ◽  
Author(s):  
Fatma Hilal Yılmaz ◽  
Mahmut Sami Tutar ◽  
Derya Arslan ◽  
Ayhan Çeri


Sign in / Sign up

Export Citation Format

Share Document