scholarly journals Literature Retrieval and Mining in Bioinformatics: State of the Art and Challenges

2012 ◽  
Vol 2012 ◽  
pp. 1-10 ◽  
Author(s):  
Andrea Manconi ◽  
Eloisa Vargiu ◽  
Giuliano Armano ◽  
Luciano Milanesi

The world has widely changed in terms of communicating, acquiring, and storing information. Hundreds of millions of people are involved in information retrieval tasks on a daily basis, in particular while using a Web search engine or searching their e-mail, making such field the dominant form of information access, overtaking traditional database-style searching. How to handle this huge amount of information has now become a challenging issue. In this paper, after recalling the main topics concerning information retrieval, we present a survey on the main works on literature retrieval and mining in bioinformatics. While claiming that information retrieval approaches are useful in bioinformatics tasks, we discuss some challenges aimed at showing the effectiveness of these approaches applied therein.

Author(s):  
R. Subhashini ◽  
V.Jawahar Senthil Kumar

The World Wide Web is a large distributed digital information space. The ability to search and retrieve information from the Web efficiently and effectively is an enabling technology for realizing its full potential. Information Retrieval (IR) plays an important role in search engines. Today’s most advanced engines use the keyword-based (“bag of words”) paradigm, which has inherent disadvantages. Organizing web search results into clusters facilitates the user’s quick browsing of search results. Traditional clustering techniques are inadequate because they do not generate clusters with highly readable names. This paper proposes an approach for web search results in clustering based on a phrase based clustering algorithm. It is an alternative to a single ordered result of search engines. This approach presents a list of clusters to the user. Experimental results verify the method’s feasibility and effectiveness.


Author(s):  
Esharenana E. Adomi

The World Wide Web (WWW) has led to the advent of the information age. With increased demand for information from various quarters, the Web has turned out to be a veritable resource. Web surfers in the early days were frustrated by the delay in finding the information they needed. The first major leap for information retrieval came from the deployment of Web search engines such as Lycos, Excite, AltaVista, etc. The rapid growth in the popularity of the Web during the past few years has led to a precipitous pronouncement of death for the online services that preceded the Web in the wired world.


Symmetry ◽  
2019 ◽  
Vol 11 (10) ◽  
pp. 1274 ◽  
Author(s):  
Md. Atiqur Rahman ◽  
Mohamed Hamada

Modern daily life activities result in a huge amount of data, which creates a big challenge for storing and communicating them. As an example, hospitals produce a huge amount of data on a daily basis, which makes a big challenge to store it in a limited storage or to communicate them through the restricted bandwidth over the Internet. Therefore, there is an increasing demand for more research in data compression and communication theory to deal with such challenges. Such research responds to the requirements of data transmission at high speed over networks. In this paper, we focus on deep analysis of the most common techniques in image compression. We present a detailed analysis of run-length, entropy and dictionary based lossless image compression algorithms with a common numeric example for a clear comparison. Following that, the state-of-the-art techniques are discussed based on some bench-marked images. Finally, we use standard metrics such as average code length (ACL), compression ratio (CR), pick signal-to-noise ratio (PSNR), efficiency, encoding time (ET) and decoding time (DT) in order to measure the performance of the state-of-the-art techniques.


2011 ◽  
Vol 1 (1) ◽  
pp. 31-44 ◽  
Author(s):  
R. Subhashini ◽  
V.Jawahar Senthil Kumar

The World Wide Web is a large distributed digital information space. The ability to search and retrieve information from the Web efficiently and effectively is an enabling technology for realizing its full potential. Information Retrieval (IR) plays an important role in search engines. Today’s most advanced engines use the keyword-based (“bag of words”) paradigm, which has inherent disadvantages. Organizing web search results into clusters facilitates the user’s quick browsing of search results. Traditional clustering techniques are inadequate because they do not generate clusters with highly readable names. This paper proposes an approach for web search results in clustering based on a phrase based clustering algorithm. It is an alternative to a single ordered result of search engines. This approach presents a list of clusters to the user. Experimental results verify the method s feasibility and effectiveness.


Author(s):  
Michael Chau ◽  
Yan Lu ◽  
Xiao Fang ◽  
Christopher C. Yang

More non-English contents are now available on the World Wide Web and the number of non-English users on the Web is increasing. While it is important to understand the Web searching behavior of these non-English users, many previous studies on Web query logs have focused on analyzing English search logs and their results may not be directly applied to other languages. In this Chapter we discuss some methods and techniques that can be used to analyze search queries in Chinese. We also show an example of applying our methods on a Chinese Web search engine. Some interesting findings are reported.


2017 ◽  
Vol 26 (06) ◽  
pp. 1730002 ◽  
Author(s):  
T. Dhiliphan Rajkumar ◽  
S. P. Raja ◽  
A. Suruliandi

Short and ambiguous queries are the major problems in search engines which lead to irrelevant information retrieval for the users’ input. The increasing nature of the information on the web also makes various difficulties for the search engine to provide the users needed results. The web search engine experience the ill effects of ambiguity, since the queries are looked at on a rational level rather than the semantic level. In this paper, for improving the performance of search engine as of the users’ interest, personalization is based on the users’ clicks and bookmarking is proposed. Modified agglomerative clustering is used in this work for clustering the results. The experimental results prove that the proposed work scores better precision, recall and F-score.


2021 ◽  
Author(s):  
◽  
Anton Findlay Angelo

<p>Institutional Repositories have been set up all over the world, and are now mainstream business for academic libraries and other organisations. The nature of the visitors or users for these repositories is not well understood, and little work has been done in analysing the data the repositories generate on their visitors. This report looks at the analytics generated by the University of Canterbury Research Repository (UCRR) through its own internal statistics and Google Analytics. There are many issues with reconciling this data, as many factors influence the accuracy of the figures, including web search engine crawlers, deep linking, and copyright trolls. This report found that there are many visitors to the UCRR, and that it is difficult, but possible to create narratives for specific items indicating how they might be used. Generalisations, however are much harder to make, and though we can see who is visiting the UCRR, we cannot really ascertain why they do. This report provides suggestions for further research on repository users, particularly at gathering qualitative data from groups identified from this quantative analytics.</p>


2021 ◽  
Author(s):  
◽  
Anton Findlay Angelo

<p>Institutional Repositories have been set up all over the world, and are now mainstream business for academic libraries and other organisations. The nature of the visitors or users for these repositories is not well understood, and little work has been done in analysing the data the repositories generate on their visitors. This report looks at the analytics generated by the University of Canterbury Research Repository (UCRR) through its own internal statistics and Google Analytics. There are many issues with reconciling this data, as many factors influence the accuracy of the figures, including web search engine crawlers, deep linking, and copyright trolls. This report found that there are many visitors to the UCRR, and that it is difficult, but possible to create narratives for specific items indicating how they might be used. Generalisations, however are much harder to make, and though we can see who is visiting the UCRR, we cannot really ascertain why they do. This report provides suggestions for further research on repository users, particularly at gathering qualitative data from groups identified from this quantative analytics.</p>


Biometrics ◽  
2017 ◽  
pp. 710-736
Author(s):  
Prashant Srivastava ◽  
Ashish Khare

The proliferation of huge amount of information has made it essential to develop systems that organize and index them for easy access. The advent of World Wide Web has provided immense opportunity to the people across the world to access and share information for different uses ranging from personal to professional. Various web mining techniques are applied to retrieve useful information as well as improvement of existing techniques of mining to search and retrieve useful information from the web. With the growth in the number of devices producing various forms of information, the amount of information is increasing exponentially. Also, these huge amount of information are being shared in the world through various means. Hence, it has become necessary to organize information in such a manner so that access to them is easy and feasible. As the amount of information is increasing rapidly, efficient indexing of information for easy access is becoming quite challenging. Hence, there is a need to search for solutions to solve this problem. The field of information retrieval attempts to solve this problem. Information retrieval is concerned with storage, organization, indexing, and retrieval of information. Information retrieval techniques incorporate several aspects of information to achieve the target of efficient indexing. Since there are several forms of information, their characteristics vary a lot from each other. Image is one such popular form of information which is shared the most among the people around the world. Also, with the presence of numerous image capturing devices, acquisition of image is no longer a difficult task. People enjoy capturing and sharing images through social network. Although image is a complex structure, it is easily understood by people across the world. Also, it has become a popular means of information sharing among people. This chapter discusses information retrieval techniques for image data. Visual Information Retrieval or Content-Based Image Retrieval (CBIR) accepts query in the form of image or image features instead of text. It is concerned with searching and retrieval of images similar to the query given in the form of images. Most of the visual information retrieval techniques are based on processing single resolution of an image. But processing of single resolution of image is not sufficient for efficient retrieval as image is a complex structure and contains varying level of details. Hence, there is a need of multiresolution processing of images. Today, it is very difficult to keep track of number of research papers based on multiresolution analysis as it is widely used for various image-based applications. Also, there are a number of multiresolution techniques available to achieve this. Multiresolution processing has one big advantage that features that are left undetected at one level get detected at another level which is not the case with single resolution analysis. We demonstrate this fact with the help of an experiment using Discrete Wavelet Transform along with the discussion of various multiresolution techniques for visual information retrieval. The experiment helps in explaining the important properties of multiresolution analysis and also provides future scope of research in this field.


Author(s):  
Iris Xie

Tim Berners-Lee wrote the initial proposal for the World Wide Web in 1989, and developed it online in 1991 by using a hypertext model (Berners-Lee, 1989, 1996). The World Wide Web was developed to allow people to collaborate on projects; it began at CERN, the European Particle Physics Laboratory in Geneva, Switzerland, and expanded across nations and disciplines. Berners-Lee (1996) defined the components of the Web: the boundless information world, the address system (URI), a network protocol (HTTP), a markup language (HTML), a body of data, and the client-server architecture of the Web. The creation in 1993 of Mosaic, a graphic Web interface that was the precursor of Netscape, enabled millions of people to easily access the Web. Since then, the increase in Web resources has been phenomenal, and Web search engines are the essential tools for navigating those Web resources.


Sign in / Sign up

Export Citation Format

Share Document