scholarly journals Rule-Based Extractive Summarization and Title Generation

Text summarying is a process by which the most important information from the source document is precisely found. It stands for the information condensed to a longer text. Text summary is broken down into two approaches: extractive summary and abstractive summary. The proposed method creates an extractive summary of a given text and generate an appropriate title for the generated summary. Extractive summary is generated through sentence selection by using Rule-based concept. Eight different features are considered to rank each sentence according to its importance. Ranking assigns a numerical measure to each sentence. After ranking, sentences that has high rank compared to others will be selected to form the summary. The frequently occurring bi-gram is selected as the title for the summary. The system performs better than existing extractive summarization techniques like Graph-based system and achieved a precision of 0.7

2018 ◽  
Vol 5 (3) ◽  
pp. 172265 ◽  
Author(s):  
Alexis R. Hernández ◽  
Carlos Gracia-Lázaro ◽  
Edgardo Brigatti ◽  
Yamir Moreno

We introduce a general framework for exploring the problem of selecting a committee of representatives with the aim of studying a networked voting rule based on a decentralized large-scale platform, which can assure a strong accountability of the elected. The results of our simulations suggest that this algorithm-based approach is able to obtain a high representativeness for relatively small committees, performing even better than a classical voting rule based on a closed list of candidates. We show that a general relation between committee size and representatives exists in the form of an inverse square root law and that the normalized committee size approximately scales with the inverse of the community size, allowing the scalability to very large populations. These findings are not strongly influenced by the different networks used to describe the individuals’ interactions, except for the presence of few individuals with very high connectivity which can have a marginal negative effect in the committee selection process.


2019 ◽  
Vol 8 (4) ◽  
pp. 1809-1814

Sentiment analysis is a technique to analyze the people opinion, attitude, sentiment and emotion towards any particular object. Sentiment analysis has the following steps to predict the opinion of a review sentences. The steps are preprocessing, feature selection, classification and sentiment prediction. Preprocessing is the main important step and it consists of many techniques. They are Stop word Removal, punctuation removal, conversion of numbers to number names. Stemming is another important preprocessing technique which is used to transform the words in text into their grammatical root form and is mainly used to improve the retrieval of the information from the internet. It is applied mainly to get strengthen the retrieval of the information. Many morphological languages have immense amount of morphological deviation in the words. It triggered vast challenges. Many algorithms exist with different techniques and has several drawbacks. The aim of this paper is to propose a rule based stemmer that is a truncating stemmer. The new stemming mechanism in this paper has brought about many morphological changes. The new rule based morphological variation removable stemming algorithm is better than the existing other algorithms such as New Porter, Paice/Lovins and Lancaster stemming algorithm


Wikipedia has recently become a popular platform for knowledge sharing and creation. However, the enormously increasing amount of editing has caused management problems with efficiency, accuracy, and convenience for Wikipedia administrators. Therefore, this study aimed to develop an intelligent agent system based on Web 3.0, the evaluation agent system (EAS), to solve these problems. The EAS is characterized by hybrid Web techniques, artificial intelligence, integration of management guidelines, retrieval of real-time information, and the transfer of cross-platform data and includes the following three systems: the testing agent, the wiki agent, and the rule-based expert system (RBES) agent. Because the RBES was central to the EAS, 29 university students were included in the study to examine the effectiveness of the RBES compared to the conventional approach to administration. The findings revealed that the RBES was better than the conventional approach in accuracy, efficiency, operation convenience, and fatigue strength.


2014 ◽  
Vol 41 (1) ◽  
pp. 56-85
Author(s):  
Delia Popescu

In this article I argue that one of the main tools that allowed the Romanian communist state to control oppositional activities, far better than many of its Eastern European neighbors, was the transformation of political opponents into petty criminals and felons. I contend that in the two decades that preceded 1989, communist Romania witnessed a pragmatic shift from hard rule (based on simply imprisoning political opponents under the category of “political detainees”) to subversive criminalization. The main operative tool for the subversive criminalization of so-called political offenses was Law 18/1968 (subtitledLaw regarding the control of the provenance of goods that have not been acquired through legal means). I argue that Law 18 was the result of two interconnected political drives. The first drive was the desire of the Ceauşescu regime to gain favor with the West by perpetuating the rhetoric launched as a result of the general amnesty for political detainees in 1964, under the Gheorghe Gheorghiu-Dej administration. The second drive was the political imperative of the Ceauşescu regime to suppress political opposition. My argument is that this transformative shift was accomplished through the development of what I call amechanism of state induced theftbacked by the deployment of a subversive legal instrument of criminalization, which was Law 18/1969. This paper analyzes the role, essence, and implications of Law 18 while supporting a theory of a strategic shift in communist policy.


2007 ◽  
Vol 30 ◽  
pp. 659-684 ◽  
Author(s):  
I. Szita ◽  
A. Lorincz

In this article we propose a method that can deal with certain combinatorial reinforcement learning tasks. We demonstrate the approach in the popular Ms. Pac-Man game. We define a set of high-level observation and action modules, from which rule-based policies are constructed automatically. In these policies, actions are temporally extended, and may work concurrently. The policy of the agent is encoded by a compact decision list. The components of the list are selected from a large pool of rules, which can be either hand-crafted or generated automatically. A suitable selection of rules is learnt by the cross-entropy method, a recent global optimization algorithm that fits our framework smoothly. Cross-entropy-optimized policies perform better than our hand-crafted policy, and reach the score of average human players. We argue that learning is successful mainly because (i) policies may apply concurrent actions and thus the policy space is sufficiently rich, (ii) the search is biased towards low-complexity policies and therefore, solutions with a compact description can be found quickly if they exist.


2019 ◽  
Author(s):  
Fransiskus Xaverius Ivan ◽  
Chee Keong Kwoh

AbstractBackgroundInfluenza A virus (IAV) poses threats to human health and life. Many individual studies have been carried out in mice to uncover the viral factors responsible for the virulence of IAV infections. Virus adaptation through serial lung-to-lung passaging and reverse genetic engineering and mutagenesis approaches have been widely used in the studies. Nonetheless, a single study may not provide enough confident about virulence factors, hence combining several studies for a meta-analysis is desired to provide better views.MethodsVirulence information of IAV infections and the corresponding virus and mouse strains were documented from literature. Using the mouse lethal dose 50, time series of weight loss or percentage of survival, the virulence of the infections was classified as avirulent or virulent for two-class problems, and as low, intermediate or high for three-class problems. On the other hand, protein sequences were decoded from the corresponding IAV genomes or reconstructed manually from other proteins according to mutations mentioned in the related literature. IAV virulence models were then learned from various datasets containing IAV proteins whose amino acids at their aligned position and the corresponding two-class or three-class virulence labels. Three proven rule-based learning approaches, i.e., OneR, JRip and PART, and additionally random forest were used for modelling, and top protein sites and synergy between protein sites were identified from the models.ResultsMore than 500 records of IAV infections in mice whose viral proteins could be retrieved were documented. The BALB/C and C57BL/6 mouse strains and the H1N1, H3N2 and H5N1 viruses dominated the infection records. PART models learned from full or subsets of datasets achieved the best performance, with moderate averaged model accuracies ranged from 65.0% to 84.4% and from 54.0% to 66.6% for two-class and three-class datasets that utilized all records of aligned IAV proteins, respectively. Their averaged accuracies were comparable or even better than the averaged accuracies of random forest models and should be preferred based on the Occam’s razor principle. Interestingly, models based on a dataset that included all IAV strains achieved a better averaged accuracy when host information was taken into account. For model interpretation, we observed that although many sites in HA were highly correlated with virulence, PART models based on sites in PB2 could compete against and were often better than PART models based on sites in HA. Moreover, PART had a high preference to include sites in PB2 when models were learned from datasets containing concatenated alignments of all IAV proteins. Several sites with a known contribution to virulence were found as the top protein sites, and site pairs that may synergistically influence virulence were also uncovered.ConclusionModelling the virulence of IAV infections is a challenging problem. Rule-based models generated using only viral proteins are useful for its advantage in interpretation, but only achieve moderate performance. Development of more advanced machine learning approaches that learn models from features extracted from both viral and host proteins must be considered for future works.


Author(s):  
Victoria A. Spaulding ◽  
Donita A. Phipps

Younger and older participants were trained to perform a computerized football task. Half of the participants received rule-based training and the remainder received color enhancements in alternating blocks. Both younger and older adults improved RT performance, with the younger participants performing about twice as fast as the older participants. The participants transferred to Novel, Cluttered and Time-Stress conditions. The effect of training type (rules better than enhancements) failed to persist during transfer. Age-related impairments of RT and overall accuracy persisted during transfer, although older participants maintained a higher primary accuracy (except for Time-Stress). Their performance plummeted during the Time-Stress, but improved across the blocks. During the subsequent baseline block, primary accuracy returned to the pre-Cluttered level and RT slightly declined. These results suggest that the older participants changed strategies under time stress, and with more practice, their performance on this complex perceptual task may increase dramatically.


BMC Genomics ◽  
2019 ◽  
Vol 20 (S9) ◽  
Author(s):  
Fransiskus Xaverius Ivan ◽  
Chee Keong Kwoh

Abstract Background Influenza A virus (IAV) poses threats to human health and life. Many individual studies have been carried out in mice to uncover the viral factors responsible for the virulence of IAV infections. Nonetheless, a single study may not provide enough confident about virulence factors, hence combining several studies for a meta-analysis is desired to provide better views. For this, we documented more than 500 records of IAV infections in mice, whose viral proteins could be retrieved and the mouse lethal dose 50 or alternatively, weight loss and/or survival data, was/were available for virulence classification. Results IAV virulence models were learned from various datasets containing aligned IAV proteins and the corresponding two virulence classes (avirulent and virulent) or three virulence classes (low, intermediate and high virulence). Three proven rule-based learning approaches, i.e., OneR, JRip and PART, and additionally random forest were used for modelling. PART models achieved the best performance, with moderate average model accuracies ranged from 65.0 to 84.4% and from 54.0 to 66.6% for the two-class and three-class problems, respectively. PART models were comparable to or even better than random forest models and should be preferred based on the Occam’s razor principle. Interestingly, the average accuracy of the models was improved when host information was taken into account. For model interpretation, we observed that although many sites in HA were highly correlated with virulence, PART models based on sites in PB2 could compete against and were often better than PART models based on sites in HA. Moreover, PART had a high preference to include sites in PB2 when models were learned from datasets containing the concatenated alignments of all IAV proteins. Several sites with a known contribution to virulence were found as the top protein sites, and site pairs that may synergistically influence virulence were also uncovered. Conclusion Modelling IAV virulence is a challenging problem. Rule-based models generated using viral proteins are useful for its advantage in interpretation, but only achieve moderate performance. Development of more advanced approaches that learn models from features extracted from both viral and host proteins shall be considered for future works.


1996 ◽  
Vol 2 (2) ◽  
pp. 95-110 ◽  
Author(s):  
JAE-HOON KIM ◽  
GIL CHANG KIM

Recently, most part-of-speech tagging approaches, such as rule-based, probabilistic and neural network approaches, have shown very promising results. In this paper, we are particularly interested in probabilistic approaches, which usually require lots of training data to get reliable probabilities. We alleviate such a restriction of probabilistic approaches by introducing a fuzzy network model to provide a method for estimating more reliable parameters of a model under a small amount of training data. Experiments with the Brown corpus show that the performance of the fuzzy network model is much better than that of the hidden Markov model under a limited amount of training data.


2013 ◽  
Vol 765-767 ◽  
pp. 1441-1445
Author(s):  
Jia Jun Cheng ◽  
Xin Zhang ◽  
Peng Yi Fan ◽  
Pei Li ◽  
Hui Wang

Chinese microblogging texts are always short and casual, which bring some troubles to the traditional sentiment classification methods based on learning. To overcome this problem, we use a rule-based approach to classify the sentiment of Chinese microblogging texts. According to the characteristics of Chinese microblogging texts, we construct a thesaurus of subjective words for it, summarize the basic semantic rules expressing emotion and propose a rule-based approach to sentiment classification of Chinese microblogging texts. Finally, we compare our approach with a SVM-based approach. Our rule-based approach achieves an accuracy of 0.865, which is better than that of SVM-based approach.


Sign in / Sign up

Export Citation Format

Share Document