scholarly journals An investigation on the evolution of diabetes data in social Q&A logs

2018 ◽  
Vol 2 (1) ◽  
pp. 37-48
Author(s):  
Yiming Zhao ◽  
Baitong Chen ◽  
Jin Zhang ◽  
Ying Ding ◽  
Jin Mao ◽  
...  

AbstractThis study investigates the evolution of diabetics’ concerns based on the analysis of terms in the Diabetes category logs on the Yahoo! Answers website. Two sets of question-and-answer (Q&A) log data were collected: one from December 2, 2005 to December 1, 2006; the other from April 1, 2013 to March 31, 2014. Network analysis and at-test were performed to analyze the differences in diabetics’ concerns between these two data sets. Community detection and topic evolution were used to reveal detailed changes in diabetics’ concerns in the examined period. Increases in average node degree and graph density imply that the vocabulary size that diabetics use to post questions decreases while the scope of questions has become more focused. The networks of key terms in the Q&A log data of 2005–2006 and 2013–2014 are significantly different according to thet-test analysis of the degree centrality and betweenness centrality. Specifically, there is a shift in diabetics’ focus in that they have become more concerned about daily life and other nonmedical issues, including diet, food, and nutrients. The recent changes and the evolution paths of diabetics’ concerns were visualized using an alluvial diagram. The food- and diet-related terms have become prominent, as deduced from the visualization results.

2021 ◽  
Vol 13 (1) ◽  
Author(s):  
Manfred Berres ◽  
Andreas U. Monsch ◽  
René Spiegel

Abstract Background The Placebo Group Simulation Approach (PGSA) aims at partially replacing randomized placebo-controlled trials (RPCTs), making use of data from historical control groups in order to decrease the needed number of study participants exposed to lengthy placebo treatment. PGSA algorithms to create virtual control groups were originally derived from mild cognitive impairment (MCI) data of the Alzheimer’s Disease Neuroimaging Initiative (ADNI) database. To produce more generalizable algorithms, we aimed to compile five different MCI databases in a heuristic manner to create a “standard control algorithm” for use in future clinical trials. Methods We compared data from two North American cohort studies (n=395 and 4328, respectively), one company-sponsored international clinical drug trial (n=831) and two convenience patient samples, one from Germany (n=726), and one from Switzerland (n=1558). Results Despite differences between the five MCI samples regarding inclusion and exclusion criteria, their baseline demographic and cognitive performance data varied less than expected. However, the five samples differed markedly with regard to their subsequent cognitive performance and clinical development: (1) MCI patients from the drug trial did not deteriorate on verbal fluency over 3 years, whereas patients in the other samples did; (2) relatively few patients from the drug trial progressed from MCI to dementia (about 10% after 4 years), in contrast to the other four samples with progression rates over 30%. Conclusion Conventional MCI criteria were insufficient to allow for the creation of well-defined and internationally comparable samples of MCI patients. More recently published criteria for MCI or “MCI due to AD” are unlikely to remedy this situation. The Alzheimer scientific community needs to agree on a standard set of neuropsychological tests including appropriate selection criteria to make MCI a scientifically more useful concept. Patient data from different sources would then be comparable, and the scientific merits of algorithm-based study designs such as the PGSA could be properly assessed.


2021 ◽  
Vol 8 (1) ◽  
Author(s):  
Hossein Ahmadvand ◽  
Fouzhan Foroutan ◽  
Mahmood Fathy

AbstractData variety is one of the most important features of Big Data. Data variety is the result of aggregating data from multiple sources and uneven distribution of data. This feature of Big Data causes high variation in the consumption of processing resources such as CPU consumption. This issue has been overlooked in previous works. To overcome the mentioned problem, in the present work, we used Dynamic Voltage and Frequency Scaling (DVFS) to reduce the energy consumption of computation. To this goal, we consider two types of deadlines as our constraint. Before applying the DVFS technique to computer nodes, we estimate the processing time and the frequency needed to meet the deadline. In the evaluation phase, we have used a set of data sets and applications. The experimental results show that our proposed approach surpasses the other scenarios in processing real datasets. Based on the experimental results in this paper, DV-DVFS can achieve up to 15% improvement in energy consumption.


Mathematics ◽  
2021 ◽  
Vol 9 (16) ◽  
pp. 1850
Author(s):  
Rashad A. R. Bantan ◽  
Farrukh Jamal ◽  
Christophe Chesneau ◽  
Mohammed Elgarhy

Unit distributions are commonly used in probability and statistics to describe useful quantities with values between 0 and 1, such as proportions, probabilities, and percentages. Some unit distributions are defined in a natural analytical manner, and the others are derived through the transformation of an existing distribution defined in a greater domain. In this article, we introduce the unit gamma/Gompertz distribution, founded on the inverse-exponential scheme and the gamma/Gompertz distribution. The gamma/Gompertz distribution is known to be a very flexible three-parameter lifetime distribution, and we aim to transpose this flexibility to the unit interval. First, we check this aspect with the analytical behavior of the primary functions. It is shown that the probability density function can be increasing, decreasing, “increasing-decreasing” and “decreasing-increasing”, with pliant asymmetric properties. On the other hand, the hazard rate function has monotonically increasing, decreasing, or constant shapes. We complete the theoretical part with some propositions on stochastic ordering, moments, quantiles, and the reliability coefficient. Practically, to estimate the model parameters from unit data, the maximum likelihood method is used. We present some simulation results to evaluate this method. Two applications using real data sets, one on trade shares and the other on flood levels, demonstrate the importance of the new model when compared to other unit models.


Author(s):  
Ahmed Masrai ◽  
James Milton ◽  
Dina Abdel Salam El-Dakhs ◽  
Heba Elmenshawy

AbstractThis study investigates the idea that knowledge of specialist subject vocabulary can make a significant and measurable impact on academic performance, separate from and additional to the impact of general and academic vocabulary knowledge. It tests the suggestion of Hyland and Tse (TESOL Quarterly, 41:235–253, 2007) that specialist vocabulary should be given more attention in teaching. Three types of vocabulary knowledge, general, academic and a specialist business vocabulary factors, are tested against GPA and a business module scores among students of business at a college in Egypt. The results show that while general vocabulary size has the greatest explanation of variance in the academic success factors, the other two factors - academic and a specialist business vocabulary - make separate and additional further contributions. The contribution to the explanation of variance made by specialist vocabulary knowledge is double that of academic vocabulary knowledge.


Author(s):  
Hezhen Hu ◽  
Wengang Zhou ◽  
Junfu Pu ◽  
Houqiang Li

Sign language recognition (SLR) is a challenging problem, involving complex manual features (i.e., hand gestures) and fine-grained non-manual features (NMFs) (i.e., facial expression, mouth shapes, etc .). Although manual features are dominant, non-manual features also play an important role in the expression of a sign word. Specifically, many sign words convey different meanings due to non-manual features, even though they share the same hand gestures. This ambiguity introduces great challenges in the recognition of sign words. To tackle the above issue, we propose a simple yet effective architecture called Global-Local Enhancement Network (GLE-Net), including two mutually promoted streams toward different crucial aspects of SLR. Of the two streams, one captures the global contextual relationship, while the other stream captures the discriminative fine-grained cues. Moreover, due to the lack of datasets explicitly focusing on this kind of feature, we introduce the first non-manual-feature-aware isolated Chinese sign language dataset (NMFs-CSL) with a total vocabulary size of 1,067 sign words in daily life. Extensive experiments on NMFs-CSL and SLR500 datasets demonstrate the effectiveness of our method.


2021 ◽  
Vol 48 (4) ◽  
pp. 307-328
Author(s):  
Dominic Farace ◽  
Hélène Prost ◽  
Antonella Zane ◽  
Birger Hjørland ◽  
◽  
...  

This article presents and discusses different kinds of data documents, including data sets, data studies, data papers and data journals. It provides descriptive and bibliometric data on different kinds of data documents and discusses the theoretical and philosophical problems by classifying documents according to the DIKW model (data documents, information documents, knowl­edge documents and wisdom documents). Data documents are, on the one hand, an established category today, even with its own data citation index (DCI). On the other hand, data documents have blurred boundaries in relation to other kinds of documents and seem sometimes to be understood from the problematic philosophical assumption that a datum can be understood as “a single, fixed truth, valid for everyone, everywhere, at all times”


2013 ◽  
Vol 16 (1) ◽  
Author(s):  
Michael Schillmeier

To assume that all things we want to describe – humans and non-humans alike – can be done so properly only in terms of 'societies', requires a contrast – a momentum of cosmopolitics – to the very abstract distinctions upon which our classical understanding of sociology and its key terms rests: 'The social' as defined in opposition to 'the non-social', 'society' in opposition to 'nature'. The concept of cosmopolitics tries to avoid such modernist strategy that A. N. Whitehead called 'bifurcation of nature' (cf. Whitehead 1978, 2000). The inventive production of contrasts names a cosmopolitical tool which does not attempt to denounce, debunk, replace or overcome abstract, exclusivist oppositions that suggest divisions as 'either…or'-relations. Rather, as the Belgian philosopher of science Isabelle Stengers stresses, 'the contrast will have to be celebrated in the manner of a new existent, adding a new dimension to the cosmos' (Stengers 2011: 513). Cosmopolitics, then, engages with 'habits we experiment with in order to become capable of new experiences' (Stengers 2001: 241) and opens up the possibility of agency of the non-expected Other, the non-normal, the non-human, the non-social, the un-common. 'The Other is the existence of a possible world', as Gilles Deleuze and Felix Guattari (1994: 17-18) have put it. It is 'the condition for our passing from one world to another. The Other (...) makes the world go by.'


2020 ◽  
Vol 13 (1) ◽  
pp. 067
Author(s):  
Christie Andre Souza ◽  
Michelle Simões Reboita

Os ciclones tropicais quando atingem ventos com intensidade igual ou superior a 119 km/h desenvolvem uma estrutura conhecida como olho em seu centro. Já os ventos mais intensos do sistema são encontrados imediatamente após o olho. Num estudo recente para os ciclones Haiyan e Haima foi levantada a questão da qualidade dos dados do Global Forecast System (GFS) em representar os ventos uma vez que os ventos máximos apareceram no olho do sistema. Diante disso, esse estudo tem como objetivo avaliar como diferentes conjuntos de dados (GFS, ERA5, ERA-Interim e CCMP) representam os ventos nesses dois ciclones tropicais. A ERA5 e o GFS mostram ventos mais intensos nos ciclones do que os outros dois conjuntos de dados. Todos, exceto o GFS, mostram claramente ventos mais fracos no olho dos ciclones.  Wind intensity of two tropical cyclones obtained by different data sets A B S T R A C TWhen the tropical cyclones reach winds with intensity equal or higher than 119 km/h, they develop a structure known as the eye at its center. The strongest winds in the system are found immediately after the eye. In a recent study for Haiyan and Haima cyclones, the question of the quality of the Global Forecast System (GFS) data in representing the winds once the maximum winds appeared in the system eye was raised. Therefore, this study aims to evaluate how different data sets (GFS, ERA5, ERA-Interim and CCMP) represent the winds in these two tropical cyclones. ERA5 and GFS show cyclones with more intense winds than the other datasets. Except the GFS, the other data clearly show weaker winds in the cyclone eye.Keywords: analyzes; cyclones; meteorology; reanalysis


2010 ◽  
Vol 23 ◽  
pp. 113-117
Author(s):  
A. Orphanou ◽  
K. Nicolaides ◽  
D. Charalambous ◽  
P. Lingis ◽  
S. C. Michaelides

Abstract. In the present study, the monthly statistical characteristics of jetlet and tropopause in relation to the development of thunderstorms over Cyprus are examined. For the needs of the study the 12:00 UTC radiosonde data obtained from the Athalassa station (33.4° E, 35.1° N) for an 11-year period, from 1997 till 2007, were employed. On the basis of this dataset, the height and the temperature of the tropopause, as well as the height, wind direction and speed of the jetlet were estimated. Additionally, the days in the above period with observed thunderstorms were selected and the aforementioned characteristics of the jetlet and tropopause were noted. The two data sets were subsequently contrasted in an attempt to identify possible relations between thunderstorm development, on the one hand, and tropopause and jetlet characteristics, on the other hand.


1986 ◽  
Vol 59 (2) ◽  
pp. 751-760
Author(s):  
Todd McLin Davis

A problem often not detected in the interpretation of survey research is the potential interaction between subgroups within the sample and aspects of the survey. Potentially interesting interactions are commonly obscured when data are analyzed using descriptive and univariate statistical procedures. This paper suggests the use of cluster analysis as a tool for interpretation of data, particularly when such data take the form of coded categories. An example of the analysis of two data sets with known properties, one random and the other contrived, is presented to illustrate the application of cluster procedures to survey research data.


Sign in / Sign up

Export Citation Format

Share Document