scholarly journals Fair: A Hadoop-based Hybrid Model for Faculty Information Retrieval System

Author(s):  
Harishchandra Dubey

In era of ever-expanding data and knowledge, we lack a centralized system that maps all the faculties to their research works. This problem has not been addressed in the past and it becomes challenging for students to connect with the right faculty of their domain. Since we have so many colleges and faculties this lies in the category of big data problem. In this paper, we present a model which works on the distributed computing environment to tackle big data. The proposed model uses apache spark as an execution engine and hive as database. The results are visualized with the help of Tableau that is connected to Apache Hive to achieve distributed computing.

Author(s):  
Ankit Shah ◽  
Mamta C. Padole

Big Data processing and analysis requires tremendous processing capability. Distributed computing brings many commodity systems under the common platform to answer the need for Big Data processing and analysis. Apache Hadoop is the most suitable set of tools for Big Data storage, processing, and analysis. But Hadoop found to be inefficient when it comes to heterogeneous set computers which have different processing capabilities. In this research, we propose the Saksham model which optimizes the processing time by efficient use of node processing capability and file management. The proposed model shows the performance improvement for Big Data processing. To achieve better performance, Saksham model uses two vital aspects of heterogeneous distributed computing: Effective block rearrangement policy and use of node processing capability. The results demonstrate that the proposed model successfully achieves better job execution time and improves data locality.


2014 ◽  
Vol 56 (5) ◽  
pp. 591-607 ◽  
Author(s):  
Rachel Kennedy ◽  
John Scriven ◽  
Magda Nenycz-Thiel

Big data is here for some and coming for many. It promises access to new knowledge along with some challenges, but let's not forget the important lessons of the past to ensure that we are advancing knowledge and making the right decisions from the data we have. In this paper, we submit that marketing's emphasis on statistical significance is misplaced, especially in the new world of big data. We include case examples to demonstrate how statistical significance is easy to find, but not necessarily important. We will also discuss the alternative route for generating robust knowledge. Specifically, we espouse the tradition pioneered by Andrew Ehrenberg of Many Sets of Data (MSoD) and descriptive models as the way to advance marketing science, and as a solid foundation for data interpretation in market research studies. We offer insights for market research practitioners and marketers alike, to ensure they are getting the best from their data for robust marketing decision-making.


Author(s):  
Ankit Shah ◽  
Mamta C. Padole

Big Data processing and analysis requires tremendous processing capability. Distributed computing brings many commodity systems under the common platform to answer the need for Big Data processing and analysis. Apache Hadoop is the most suitable set of tools for Big Data storage, processing, and analysis. But Hadoop found to be inefficient when it comes to heterogeneous set computers which have different processing capabilities. In this research, we propose the Saksham model which optimizes the processing time by efficient use of node processing capability and file management. The proposed model shows the performance improvement for Big Data processing. To achieve better performance, Saksham model uses two vital aspects of heterogeneous distributed computing: Effective block rearrangement policy and use of node processing capability. The results demonstrate that the proposed model successfully achieves better job execution time and improves data locality.


Commonwealth ◽  
2017 ◽  
Vol 19 (1) ◽  
Author(s):  
John Arway

The challenges of including factual information in public policy and political discussions are many. The difficulties of including scientific facts in these debates can often be frustrating for scientists, politicians and policymakers alike. At times it seems that discussions involve different languages or dialects such that it becomes a challenge to even understand one another’s position. Oftentimes difference of opinion leads to laws and regulations that are tilted to the left or the right. The collaborative balancing to insure public and natural resource interests are protected ends up being accomplished through extensive litigation in the courts. In this article, the author discusses the history of environmental balancing during the past three decades from the perspective of a field biologist who has used the strength of our policies, laws and regulations to fight for the protection of our Commonwealth’s aquatic resources. For the past 7 years, the author has taken over the reins of “the most powerful environmental agency in Pennsylvania” and charted a course using science to properly represent natural resource interests in public policy and political deliberations.


Author(s):  
. Monika ◽  
Pardeep Kumar ◽  
Sanjay Tyagi

In Cloud computing environment QoS i.e. Quality-of-Service and cost is the key element that to be take care of. As, today in the era of big data, the data must be handled properly while satisfying the request. In such case, while handling request of large data or for scientific applications request, flow of information must be sustained. In this paper, a brief introduction of workflow scheduling is given and also a detailed survey of various scheduling algorithms is performed using various parameter.


1996 ◽  
Vol 35 (4I) ◽  
pp. 399-417 ◽  
Author(s):  
John W. Mellor

The right to the flow of income from water is vigorously pursued, protected, and fought over in any arid part of the world. Pakistan is of course no exception. Reform of irrigation institutions necessarily changes the rights to water, whether it be those of farmers, government, or government functionaries. Those perceived rights may be explicit and broadly accepted, or simply takings that are not even considered legitimate. Nevertheless they will be fought over. Pakistan has a long history of proposals for irrigation reform, little or none being implemented, except as isolated pilot projects. Thus, to propose major changes in irrigation institutions must be clearly shown to have major benefits to justify the hard battles that must be fought and the goodwill of those who might win those battles for reform. Proponents of irrigation institution reform have always argued the necessity of the reforms and the large gains to be achieved. Perhaps, however, those arguments have not been convincing. This paper will briefly outline the failed attempts at irrigation reform to provide an element of reality to the discussion. It will then proceed to make the case of the urgency of reform in a somewhat different manner to the past. Finally, current major reform proposals will be presented. This paper approaches justification of irrigation reform by focusing on the agricultural growth rate. It does so because that is the critical variable influencing poverty rates and is a significant determinant of over-all economic growth rates. The paper decomposes growth rates and suggests a residual effect of deterioration of the irrigation system that is large and calls for policy and institutional reform. The data are notional, suggesting the usefulness of the approach and paves the way for more detailed empirical analysis and enquiry for the future.


2020 ◽  
Author(s):  
Anusha Ampavathi ◽  
Vijaya Saradhi T

UNSTRUCTURED Big data and its approaches are generally helpful for healthcare and biomedical sectors for predicting the disease. For trivial symptoms, the difficulty is to meet the doctors at any time in the hospital. Thus, big data provides essential data regarding the diseases on the basis of the patient’s symptoms. For several medical organizations, disease prediction is important for making the best feasible health care decisions. Conversely, the conventional medical care model offers input as structured that requires more accurate and consistent prediction. This paper is planned to develop the multi-disease prediction using the improvised deep learning concept. Here, the different datasets pertain to “Diabetes, Hepatitis, lung cancer, liver tumor, heart disease, Parkinson’s disease, and Alzheimer’s disease”, from the benchmark UCI repository is gathered for conducting the experiment. The proposed model involves three phases (a) Data normalization (b) Weighted normalized feature extraction, and (c) prediction. Initially, the dataset is normalized in order to make the attribute's range at a certain level. Further, weighted feature extraction is performed, in which a weight function is multiplied with each attribute value for making large scale deviation. Here, the weight function is optimized using the combination of two meta-heuristic algorithms termed as Jaya Algorithm-based Multi-Verse Optimization algorithm (JA-MVO). The optimally extracted features are subjected to the hybrid deep learning algorithms like “Deep Belief Network (DBN) and Recurrent Neural Network (RNN)”. As a modification to hybrid deep learning architecture, the weight of both DBN and RNN is optimized using the same hybrid optimization algorithm. Further, the comparative evaluation of the proposed prediction over the existing models certifies its effectiveness through various performance measures.


2021 ◽  
Vol 11 (13) ◽  
pp. 6047
Author(s):  
Soheil Rezaee ◽  
Abolghasem Sadeghi-Niaraki ◽  
Maryam Shakeri ◽  
Soo-Mi Choi

A lack of required data resources is one of the challenges of accepting the Augmented Reality (AR) to provide the right services to the users, whereas the amount of spatial information produced by people is increasing daily. This research aims to design a personalized AR that is based on a tourist system that retrieves the big data according to the users’ demographic contexts in order to enrich the AR data source in tourism. This research is conducted in two main steps. First, the type of the tourist attraction where the users interest is predicted according to the user demographic contexts, which include age, gender, and education level, by using a machine learning method. Second, the correct data for the user are extracted from the big data by considering time, distance, popularity, and the neighborhood of the tourist places, by using the VIKOR and SWAR decision making methods. By about 6%, the results show better performance of the decision tree by predicting the type of tourist attraction, when compared to the SVM method. In addition, the results of the user study of the system show the overall satisfaction of the participants in terms of the ease-of-use, which is about 55%, and in terms of the systems usefulness, about 56%.


Sign in / Sign up

Export Citation Format

Share Document