XWRAPComposer

Author(s):  
Ling Liu ◽  
Jianjun Zhang ◽  
Wei Han ◽  
Calton Pu ◽  
James Caverlee ◽  
...  

We present a service-oriented architecture and a set of techniques for developing wrapper code generators, including the methodology of designing an effective wrapper program construction facility and a concrete implementation, called XWRAPComposer. Our wrapper generation framework has two unique design goals. First, we explicitly separate tasks of building wrappers that are specific to a Web service from the tasks that are repetitive for any service, thus the code can be generated as a wrapper library component and reused automatically by the wrapper generator system. Second, we use inductive learning algorithms that derive information flow and data extraction patterns by reasoning about sample pages or sample specifications. More importantly, we design a declarative rule-based script language for multi-page information extraction, encouraging a clean separation of the information extraction semantics from the information flow control and execution logic of wrapper programs. We implement these design principles with the development of the XWRAPComposer toolkit, which can semi-automatically generate WSDL-enabled wrapper programs. We illustrate the problems and challenges of multi-page data extraction in the context of bioinformatics applications and evaluate the design and development of XWRAPComposer through our experiences of integrating various BLAST services.

Author(s):  
Ling Liu ◽  
Jianjun Zhang ◽  
Wei Han ◽  
Calton Pu ◽  
James Caverlee ◽  
...  

We present a service-oriented architecture and a set of techniques for developing wrapper code generators, including the methodology of designing an effective wrapper program construction facility and a concrete implementation, called XWRAPComposer. Our wrapper generation framework has two unique design goals. First, we explicitly separate tasks of building wrappers that are specific to a Web service from the tasks that are repetitive for any service, thus the code can be generated as a wrapper library component and reused automatically by the wrapper generator system. Second, we use inductive learning algorithms that derive information flow and data extraction patterns by reasoning about sample pages or sample specifications. More importantly, we design a declarative rule-based script language for multi-page information extraction, encouraging a clean separation of the information extraction semantics from the information flow control and execution logic of wrapper programs. We implement these design principles with the development of the XWRAPComposer toolkit, which can semi-automatically generate WSDL-enabled wrapper programs. We illustrate the problems and challenges of multi-page data extraction in the context of bioinformatics applications and evaluate the design and development of XWRAPComposer through our experiences of integrating various BLAST services.


2013 ◽  
Vol 336-338 ◽  
pp. 2348-2353 ◽  
Author(s):  
Bo Yu ◽  
Lin Yang ◽  
Shu Hui Chen ◽  
Lin Ru Ma

Service computing facilitates resource sharing and business collaboration for cross-domain partiers by universal service description and discovery. The multi-domain nature of service oriented environments introduces challenging security issues, especially with regard to information flow control which controls the flow of privacy resources. This paper presents a state-of-the-art review of information flow control technology in service oriented environments. We review the research background and classifications of information flow control, and discuss architecture and existing technologies of both centralized and distributed information flow approach. We outline the features, advantages and limitations of existing information flow control approaches. The analysis results show that the research of collaboration and dynamic nature are important, but insufficient to secure information flow.


2021 ◽  
Author(s):  
Cristina Paniagua ◽  
Jerker Delsing

Abstract The new Industry 4.0 approach contributes to addressing evolving industrial requirements, which are continuously fueled by changing market demands. This situation leads to growing complexity and considerable increases in development and maintenance costs. A significant portion of engineering time is dedicated to the integration and interconnection of heterogeneous components. The solution for interoperability issues and the reduction in the associated engineering time are thus key tasks for increasing productivity and efficiency. Therefore, this paper provides an engineering approach to create interoperability among heterogeneous systems in Service Oriented Architecture (SOA) based environments by means of generating an autonomous consumer interface code at runtime. This paper aims to present a novel interoperability solution. The proposed approach makes use of service interface descriptions to dynamically instantiate a new autonomously generated interface that solves service mismatches between a provider and a consumer. This paper includes the definition of the consumer interface generator system, as well as the benefits and challenges associated with the autonomous generation and deployment of a consumer interface code at runtime. To illustrate the potential of this approach, a prototype of the system, which shows positive results, is implemented and tested.


2013 ◽  
Vol 7 (2) ◽  
pp. 574-579 ◽  
Author(s):  
Dr Sunitha Abburu ◽  
G. Suresh Babu

Day by day the volume of information availability in the web is growing significantly. There are several data structures for information available in the web such as structured, semi-structured and unstructured. Majority of information in the web is presented in web pages. The information presented in web pages is semi-structured.  But the information required for a context are scattered in different web documents. It is difficult to analyze the large volumes of semi-structured information presented in the web pages and to make decisions based on the analysis. The current research work proposed a frame work for a system that extracts information from various sources and prepares reports based on the knowledge built from the analysis. This simplifies  data extraction, data consolidation, data analysis and decision making based on the information presented in the web pages.The proposed frame work integrates web crawling, information extraction and data mining technologies for better information analysis that helps in effective decision making.   It enables people and organizations to extract information from various sourses of web and to make an effective analysis on the extracted data for effective decision making.  The proposed frame work is applicable for any application domain. Manufacturing,sales,tourisum,e-learning are various application to menction few.The frame work is implemetnted and tested for the effectiveness of the proposed system and the results are promising.


Sign in / Sign up

Export Citation Format

Share Document