scholarly journals XML and Semantics

Author(s):  
Mohammad Moradi ◽  
MohammadReza Keyvanpour

Since the early days of introducing eXtensible Markup Language (XML), owing to its expressive capabilities and flexibilities, it became the defacto standard for representing, storing, and interchanging data on the Web. Such features have made XML one of the building blocks of the Semantic Web. From another viewpoint, since XML documents could be considered from content, structural, and semantic aspects, leveraging their semantics is very useful and applicable in different domains. However, XML does not by itself introduce any built-in mechanisms for governing semantics. For this reason, many studies have been conducted on the representation of semantics within/from XML documents. This paper studies and discusses different aspects of the mentioned topic in the form of an overview with an emphasis on the state of semantics in XML and its presentation methods.

Author(s):  
Kevin R. Parker

Before understanding the Semantic Web and its associated benefits, one must first be somewhat familiar with the enabling technologies upon which the Semantic Web is based. The extensible markup language (XML), uniform resource identifiers (URIs), resource definition framework (RDF), ontologies, and intelligent agents are all key tithe realization of the Semantic Web. Understanding these key technologies gives readers a firm foundation before progressing to subsequent chapters. This chapter provides a broad overview of each technology, and readers new to these technologies are provided with references to more detailed explanations.


Author(s):  
Adélia Gouveia ◽  
Jorge Cardoso

The World Wide Web (WWW) emerged in 1989, developed by Tim Berners-Lee who proposed to build a system for sharing information among physicists of the CERN (Conseil Européen pour la Recherche Nucléaire), the world’s largest particle physics laboratory. Currently, the WWW is primarily composed of documents written in HTML (hyper text markup language), a language that is useful for visual presentation (Cardoso & Sheth, 2005). HTML is a set of “markup” symbols contained in a Web page intended for display on a Web browser. Most of the information on the Web is designed only for human consumption. Humans can read Web pages and understand them, but their inherent meaning is not shown in a way that allows their interpretation by computers (Cardoso & Sheth, 2006). Since the visual Web does not allow computers to understand the meaning of Web pages (Cardoso, 2007), the W3C (World Wide Web Consortium) started to work on a concept of the Semantic Web with the objective of developing approaches and solutions for data integration and interoperability purpose. The goal was to develop ways to allow computers to understand Web information. The aim of this chapter is to present the Web ontology language (OWL) which can be used to develop Semantic Web applications that understand information and data on the Web. This language was proposed by the W3C and was designed for publishing, sharing data and automating data understood by computers using ontologies. To fully comprehend OWL we need first to study its origin and the basic blocks of the language. Therefore, we will start by briefly introducing XML (extensible markup language), RDF (resource description framework), and RDF Schema (RDFS). These concepts are important since OWL is written in XML and is an extension of RDF and RDFS.


Author(s):  
Claus Huitfeldt

This chapter describes how digital critical edition supposes a mastery of markup systems, providing an overview in the form of an inventory of standards, and of markup, presentation, and archiving techniques. It discusses the state of the art while focusing on key architectures and techniques considered as the basis of digital critical edition. The chapter introduces some aspects of markup technology that are particularly relevant to textual scholarship, such as the Extensible Markup Language (XML) and the Text Encoding Initiative (TEI), and considers some of their limitations, possibilities, and future potential. Since there is no need to be conversant with all aspects and details of the markup technology, most of what is covered here is of a general nature, albeit focusing on issues assumed to be of particular relevance for textual scholarship.


Author(s):  
Domenic Denicola

XML's steady descent into obscurity has become more and more apparent over the last few years. Developers, tool vendors, and browser implementers have all embraced HTML as the web's markup language, built on a substrate of JavaScript. Nothing epitomizes this shift more than the recent rise of web components: instead of standards committees dreaming up domain-specific XML vocabularies and hoping one day browsers would incorporate them, web components and the extensible web principles they embody allow authors to empower HTML with the same abilities XML once promised. The HTML of today is a truly extensible markup language. Where XML failed in this mission, both historically and practically, the web ecosystem routed around the damage of XML's influence by making HTML better suited for extensibility than ever before.


2011 ◽  
pp. 879-899
Author(s):  
Laura Irina Rusu ◽  
Wenny Rahayu ◽  
David Taniar

This chapter presents some of the existing mining techniques for extracting association rules out of XML documents in the context of rapid changes in the Web knowledge discovery area. The initiative of this study was driven by the fast emergence of XML (eXtensible Markup Language) as a standard language for representing semistructured data and as a new standard of exchanging information between different applications. The data exchanged as XML documents become richer and richer every day, so the necessity to not only store these large volumes of XML data for later use, but to mine them as well to discover interesting information has became obvious. The hidden knowledge can be used in various ways, for example, to decide on a business issue or to make predictions about future e-customer behaviour in a Web application. One type of knowledge that can be discovered in a collection of XML documents relates to association rules between parts of the document, and this chapter presents some of the top techniques for extracting them.


Author(s):  
Richard Crowder ◽  
Yee-Wie Sim

Organisations are increasingly information intensive; hence providing access to data that is trapped in various proprietary forms including catalogues, databases, human resource systems and internally generated documents is now becoming a significant and challenging task. The authors have undertaken research into approaches to capture relevant knowledge from legacy documents. This is achieved by converting the legacy documents to XML, (eXtensible Markup Language), documents where the output is semantically tagged. Once in an XML form, the data can be easily transformed. This paper describes the development of tools to automate the process of converting legacy documents to XML documents. The purpose of this work is improve the efficiency and reliability of Expertise Finder suitable for use within an engineering design environment. We will also show that by querying the resultant XML versions of legacy documents provides better results than a basic text search over the identical documents when applied used within an Expertise Finder.


Author(s):  
Badya Al-Hamadani ◽  
Joan Lu

The eXtensible Markup Language (XML) is a World Wide Web Consortium (W3C) recommendation which has widely been used in both commerce and research. As the importance of XML documents increase, the need to deal with these documents increases as well. This chapter illustrates the methodology that has been used throughout the research, discussing all its parts and how these parts were adopted in the research.


Author(s):  
Laura Irina Rusu ◽  
Wenny Rahayu ◽  
David Taniar

This chapter presents some of the existing mining techniques for extracting association rules out of XML documents in the context of rapid changes in the Web knowledge discovery area. The initiative of this study was driven by the fast emergence of XML (eXtensible Markup Language) as a standard language for representing semistructured data and as a new standard of exchanging information between different applications. The data exchanged as XML documents become richer and richer every day, so the necessity to not only store these large volumes of XML data for later use, but to mine them as well to discover interesting information has became obvious. The hidden knowledge can be used in various ways, for example, to decide on a business issue or to make predictions about future e-customer behaviour in a Web application. One type of knowledge that can be discovered in a collection of XML documents relates to association rules between parts of the document, and this chapter presents some of the top techniques for extracting them.


2007 ◽  
pp. 79-103 ◽  
Author(s):  
Laura Irina Rusu ◽  
Wenny Rahayu ◽  
David Taniar

This chapter presents some of the existing mining techniques for extracting association rules out of XML documents, in the context of rapid changes in the Web knowledge discovery area. The initiative of this study was driven by the fast emergence of XML (eXtensible Markup Language) as a standard language for representing semi-structured data and as a new standard of exchanging information between different applications. The data exchanged as XML documents becomes every day richer and richer, so the necessity to not only store these large volume of XML data for later use, but to mine them as well, to discover interesting information, has became obvious. The hidden knowledge can be used in various ways, for example to decide on a business issue or to make predictions about future e-customer behaviour in a web-application. One type of knowledge which can be discovered in a collection of XML documents relates to association rules between parts of the document, and this chapter presents some of the top techniques for extracting them.


2012 ◽  
Author(s):  
Ren Hui Gong ◽  
Ziv Yaniv

The Insight Segmentation and Registration Toolkit (ITK) previously provided a framework for parsing Extensible Markup Language (XML) documents using the Simple API for XML (SAX) framework. While this programming model is memory efficient, it places most of the implementation burden on the user. We provide an implementation of the Document Object Model (DOM) framework for parsing XML documents. Using this model, user code is greatly simplified, shifting most of the implementation burden from the user to the framework. The provided implementation consists of two tiers. The lower level tier provides functionality for parsing XML documents and loading the tree structure into memory. It then allows the user to query and retrieve specific entries. The upper tier uses this functionality to provide an interface for mimicking a serialization and de-serialization mechanism for ITK objects. The implementation described in this document was incorporated into ITK as part of release 4.2.


Sign in / Sign up

Export Citation Format

Share Document