scholarly journals Selection of the Optimal Number of Topics for LDA Topic Model—Taking Patent Policy Analysis as an Example

Entropy ◽  
2021 ◽  
Vol 23 (10) ◽  
pp. 1301
Author(s):  
Jingxian Gan ◽  
Yong Qi

This study constructs a comprehensive index to effectively judge the optimal number of topics in the LDA topic model. Based on the requirements for selecting the number of topics, a comprehensive judgment index of perplexity, isolation, stability, and coincidence is constructed to select the number of topics. This method provides four advantages to selecting the optimal number of topics: (1) good predictive ability, (2) high isolation between topics, (3) no duplicate topics, and (4) repeatability. First, we use three general datasets to compare our proposed method with existing methods, and the results show that the optimal topic number selection method has better selection results. Then, we collected the patent policies of various provinces and cities in China (excluding Hong Kong, Macao, and Taiwan) as datasets. By using the optimal topic number selection method proposed in this study, we can classify patent policies well.

1981 ◽  
Vol 29 ◽  
pp. 1-9
Author(s):  
George J. Graham

The purpose of this course is to introduce a new framework linking the humanities to public policy analysis as pursued in the government and the academy. Current efforts to link the particular contributions from the humanities to problems of public policy choice are often narrow either in terms of their perspective on the humanities or in terms of their selection of the possible means of influencing policy choice. Sometimes a single text from one of the humanities disciplines is selected to apply to a particular issue. At other times, arguments about the ethical dimensions of a single policy issue often are pursued with a single — or sometimes, no — point of access to the policy process in mind.


Sensors ◽  
2021 ◽  
Vol 21 (10) ◽  
pp. 3311
Author(s):  
Riccardo Ballarini ◽  
Marco Ghislieri ◽  
Marco Knaflitz ◽  
Valentina Agostini

In motor control studies, the 90% thresholding of variance accounted for (VAF) is the classical way of selecting the number of muscle synergies expressed during a motor task. However, the adoption of an arbitrary cut-off has evident drawbacks. The aim of this work is to describe and validate an algorithm for choosing the optimal number of muscle synergies (ChoOSyn), which can overcome the limitations of VAF-based methods. The proposed algorithm is built considering the following principles: (1) muscle synergies should be highly consistent during the various motor task epochs (i.e., remaining stable in time), (2) muscle synergies should constitute a base with low intra-level similarity (i.e., to obtain information-rich synergies, avoiding redundancy). The algorithm performances were evaluated against traditional approaches (threshold-VAF at 90% and 95%, elbow-VAF and plateau-VAF), using both a simulated dataset and a real dataset of 20 subjects. The performance evaluation was carried out by analyzing muscle synergies extracted from surface electromyographic (sEMG) signals collected during walking tasks lasting 5 min. On the simulated dataset, ChoOSyn showed comparable performances compared to VAF-based methods, while, in the real dataset, it clearly outperformed the other methods, in terms of the fraction of correct classifications, mean error (ME), and root mean square error (RMSE). The proposed approach may be beneficial to standardize the selection of the number of muscle synergies between different research laboratories, independent of arbitrary thresholds.


2014 ◽  
Vol 1061-1062 ◽  
pp. 974-977
Author(s):  
Shi Hua Liu ◽  
Xian Gang Liu ◽  
Zhi Jian Sun

A skywave radar adaptive frequency selection method based on the preliminary criterion and the weighted criterion is presented. In this method, according to the various operational tasks, the frequency selection criterion is divided into the preliminary criterion and the weighted criterion based on the characteristic of the targets. The adaptive frequency selection of the skywave radar is achieved by the weighted computed of the frequency selection criterion. The feasibility and availability is demonstrated by an example.


2013 ◽  

Over the last 30 years, the Inforum approach to macro modelling has been shared by economists worldwide. Researchers have focussed much of their efforts to developing a linked system of international interindustry models with a consistent methodology. A world-wide network of research associates use the same methods and software obtaining comparable results. The XXth Inforum World Conference was held in Florence in September 2012 and this book contains a selection of papers presented during that Conference. All these contributions are aimed at policymakers, stakeholders, and applied economists. Some papers are devoted to specific topics (total factor productivity, energy issues, external linkages, demographic changes) and some others are oriented to macro model building and simulations.


2015 ◽  
Vol 1 (311) ◽  
Author(s):  
Piotr Tarka

Abstract: The objective article is the comparative analysis of Likert rating scale based on the following range of response categories, i.e. 5, 7, 9 and 11 in context of the appropriate process of factors extraction in exploratory factor analysis (EFA). The problem which is being addressed in article relates primarily to the methodological aspects, both in selection of the optimal number of response categories of the measured items (constituting the Likert scale) and identification of possible changes, differences or similarities associated (as a result of the impact of four types of scales) with extraction and determination the appropriate number of factors in EFA model.Keywords: Exploratory factor analysis, Likert scale, experiment research, marketing


2006 ◽  
Vol 24 (4) ◽  
pp. 349-365 ◽  
Author(s):  
C. M. Tang ◽  
C. W. Y. Wong ◽  
A. Y. T. Leung ◽  
K. C. Lam

2018 ◽  
Vol 26 (5) ◽  
pp. 594-604 ◽  
Author(s):  
Javier Astudillo ◽  
Klaus Detterbeck

In many Western democracies, political parties have started to open to members the selection of their leaders. While most studies focus on the introduction of this new selection method, its subsequent practice is still understudied. The article contributes to our still limited knowledge of this process by looking at two multilevel countries, Germany and Spain, where the mainstream parties have sometimes organized membership ballots, especially at the regional level, for leadership selection. Thanks to two original databases on party conferences and membership ballots, the article analyzes the background of this process and reviews the most common explanations offered by the literature. It shows that they are not held when parties want to regain power, or party chairs seek their nomination, as commonly believed, but when there are intraparty leadership disputes.


2018 ◽  
Vol 63 (5) ◽  
pp. 529-535 ◽  
Author(s):  
Tobias Heimpold ◽  
Frank Reifegerste ◽  
Stefan Drechsel ◽  
Jens Lienig

AbstractHyperspectral imaging (HSI) has become a sophisticated technique in modern applications such as food analyses, recycling technology, medicine, pharmacy and forensic science. It allows one to analyse both spatial and spectral information from an object. But hyperspectral cameras are still expensive due to their extended wavelength range. The development of new light-emitting diodes (LED) in the recent past enables another approach to HSI using a monochrome camera in combination with a LED-based illumination. However, such a system has a lower spectral resolution. Additionally, the growing supply of LED on the market complicates the selection of LED. In this paper, we propose a new time efficient selection method for the design process of an illumination. It chooses an optimised LED combination from an existing database to match a predefined spectral power distribution. Therefore, an algorithm is used to evaluate various LED combinations. Furthermore, the method considers the spectral behaviour of each LED in dependence of forward current and temperature of the solder point. Our method has already shown promise during the selection process for even spectral distributions which is demonstrated in the study. Additionally, we will show its potential for HSI illuminations.


2021 ◽  
pp. 1-16
Author(s):  
Aikaterini Karanikola ◽  
Charalampos M. Liapis ◽  
Sotiris Kotsiantis

In short, clustering is the process of partitioning a given set of objects into groups containing highly related instances. This relation is determined by a specific distance metric with which the intra-cluster similarity is estimated. Finding an optimal number of such partitions is usually the key step in the entire process, yet a rather difficult one. Selecting an unsuitable number of clusters might lead to incorrect conclusions and, consequently, to wrong decisions: the term “optimal” is quite ambiguous. Furthermore, various inherent characteristics of the datasets, such as clusters that overlap or clusters containing subclusters, will most often increase the level of difficulty of the task. Thus, the methods used to detect similarities and the parameter selection of the partition algorithm have a major impact on the quality of the groups and the identification of their optimal number. Given that each dataset constitutes a rather distinct case, validity indices are indicators introduced to address the problem of selecting such an optimal number of clusters. In this work, an extensive set of well-known validity indices, based on the approach of the so-called relative criteria, are examined comparatively. A total of 26 cluster validation measures were investigated in two distinct case studies: one in real-world and one in artificially generated data. To ensure a certain degree of difficulty, both real-world and generated data were selected to exhibit variations and inhomogeneity. Each of the indices is being deployed under the schemes of 9 different clustering methods, which incorporate 5 different distance metrics. All results are presented in various explanatory forms.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
E.M.A.C. Ekanayake ◽  
Geoffrey Qiping Shen ◽  
Mohan Kumaraswamy ◽  
Emmanuel Kingsford Owusu

PurposeIndustrialized construction (IC) has been recognized as a game-changing approach in Hong Kong (HK). However, the increasing risks of disruptions in IC supply chains (SCs) raise SC vulnerability levels, prompting attention to developing supply chain resilience (SCR). Since SCR is only attainable through overcoming critical supply chain vulnerabilities (CSCV) with enhanced SC capabilities, this study first aimed to determine the most CSCV of ICSCs by addressing this current research gap and practical need.Design/methodology/approachDrawing on SCV factors identified from a precursor literature review, an empirical study of IC in HK was conducted using a questionnaire survey and interviews with industry experts. Focussed significance analysis of the data collected through questionnaire survey enabled the selection of 26 CSCV as appropriate to IC. Next, factor analysis was conducted, enabling the grouping of these CSCV under five components. The results were verified and reinforced by interview findings.FindingsThe results revealed 26 CSCV pertinent to resilient ICSCs in HK with five underlying components: economic, technological, procedural, organizational and production-based vulnerabilities. Loss of skilled labour is the most critical vulnerability, whereas organizational SCV is the most critical component identified.Originality/valueFindings of this study would motivate IC project professionals to appreciate and address the CSCV in the context of five components and thereby develop adequate specific capabilities to successfully withstand these CSCV. This should trigger future studies to map CSCV with appropriate capabilities in developing an envisaged powerful assessment model for evaluating the SCR in IC in HK.


Sign in / Sign up

Export Citation Format

Share Document