scholarly journals A Bibliometric Analysis of 34,692 Publications on Thyroid Cancer by Machine Learning: How Much Has Been Done in the Past Three Decades?

2021 ◽  
Vol 11 ◽  
Author(s):  
Zeyu Zhang ◽  
Lei Yao ◽  
Wenlong Wang ◽  
Bo Jiang ◽  
Fada Xia ◽  
...  

IntroductionThyroid cancer (TC) is the most common neck malignancy. However, a large number of publications of TC have not been well summarized and discussed with more comprehensive methods. The purpose of this bibliometric study is to summarize scientific publications during the past three decades in the field of TC using a machine learning method.Material and MethodsScientific publications focusing on TC from 1990 to 2020 were searched in PubMed using the MeSH term “thyroid neoplasms”. Full associated data were downloaded in the format of PubMed, and extracted in the R platform. Latent Dirichlet allocation (LDA) was adopted to identify the research topics from the abstract of each publication using Python.ResultsA total of 34,692 publications related to TC from the last three decades were found and included in this study with an average of 1,119.1 publications per year. Clinical studies and experimental studies shared the most proportion of publications, while the proportion of clinical trials remained at a relatively small level (5.87% as the highest in 2004). Thyroidectomy was the lead MeSH term, followed by prognosis, differential diagnosis, and fine-needle biopsy. The LDA analyses showed the study topics were divided into four clusters, including treatment management, basic research, diagnosis research, epidemiology, and cancer risk. However, a relatively weak connection was shown between treatment managements and basic researches. Top 10 most cited publications in recent years particularly highlighted the applications of active surveillance in TC.ConclusionThyroidectomy, differential diagnosis, genomic analysis, active surveillance are the most concerning topics in TC researches. Although the BRAF-targeted therapy is under development with promising results, there is still an urgent need for conversions from basic studies to clinical practice.

2021 ◽  
Vol 11 ◽  
Author(s):  
Zeyu Zhang ◽  
Zhiming Wang ◽  
Yun Huang

IntroductionCholangiocarcinoma (CCA) is the second most common hepatic malignancy. Progress and developments have also been made in the field of CCA management along with increasing scientific publications during the past decades, which reflect topics of general interest and suggest the future direction of studies. The purpose of this bibliometric study is to summarize scientific publications during the past 25 years in the field of CCA using a machine learning method.Material and MethodsScientific publications focusing on CCA from 1995 to 2019 were searched in PubMed using the MeSH term “cholangiocarcinoma.” Full associated data were downloaded in the format of PubMed and extracted in the R platform. Latent Dirichlet allocation (LDA) was adopted to identify the research topics from the abstract of each publication using Python.ResultsA total of 8,276 publications related to CCA from the last 25 years were found and included in this study. The most type of publications remained little changed, while the proportion of clinical trials remained relatively low (7.24% as the highest) and, more significantly, with a further downward trend during the recent years (1.42% in 2019). Neoplasm staging, hepatectomy, and survival rate were the most concerning terms among those who are diagnosis-related, treatment-related, and prognosis-related. The LDA analyses showed chemotherapy, hepatectomy, and stent as the highly concerned research topics of CCA treatment. Meanwhile, conversions from basic studies to clinical therapies were suggested by a poor connection between clusters of treatment management and basic research.ConclusionThe number of publications of CCA has increased rapidly during the past 25 years. Survival analysis, differential diagnosis, and microRNA expression are the most concerned topics in CCA studies. Besides, there is an urgent need for high-quality clinical trials and conversions from basic studies to clinical therapies.


2020 ◽  
Vol 13 ◽  
pp. 175628482093459
Author(s):  
Kangtao Wang ◽  
Chenzhe Feng ◽  
Ming Li ◽  
Qian Pei ◽  
Yuqiang Li ◽  
...  

Background and Aims: The aim of this study was to analyse the landscape of publications on rectal cancer (RC) over the past 25 years by machine learning and semantic analysis. Methods: Publications indexed in PubMed under the Medical Subject Headings (MeSH) term ‘Rectal Neoplasms’ from 1994 to 2018 were downloaded in September 2019. R and Python were used to extract publication date, MeSH terms and abstract from the metadata of each publication for bibliometric assessment. Latent Dirichlet allocation was applied to analyse the text from the articles’ abstracts to identify more specific research topics. Louvain algorithm was used to establish a topic network resulting in identifying the relationship between the topics. Results: A total of 23,492 papers published were identified and analysed in this study. The changes of research focus were analysed by the changing of MeSH terms. Studied contents extracted from the publications were divided into five areas, including surgical intervention, radiotherapy and chemotherapy intervention, clinical case management, epidemiology and cancer risk as well as prognosis studies. Conclusions: The number of publications indexed on RC has expanded rapidly over the past 25 years. Studies on RC have mainly focused on five areas. However, studies on basic research, postoperative quality of life and cost-effective research were relatively lacking. It is predicted that basic research, inflammation and some other research fields might become the potential hotspots in the future.


Author(s):  
Chan Li ◽  
Zhaoya Liu ◽  
Ruizheng Shi

Myocardial ischemia is the major cause of death worldwide, and reperfusion is the standard intervention for myocardial ischemia. However, reperfusion may cause additional damage, known as myocardial reperfusion injury, for which there is still no effective therapy. This study aims to analyze the landscape of researches concerning myocardial reperfusion injury over the past three decades by machine learning. PubMed was searched for publications from 1990 to 2020 indexed under the Medical Subject Headings (MeSH) term “myocardial reperfusion injury” on 13 April 2021. MeSH analysis and Latent Dirichlet allocation (LDA) analyses were applied to reveal research hotspots. In total, 14,822 publications were collected and analyzed in this study. MeSH analyses revealed that time factors and apoptosis were the leading terms of the pathogenesis and treatment of myocardial reperfusion injury, respectively. In LDA analyses, research topics were classified into three clusters. Complex correlations were observed between topics of different clusters, and the prognosis is the most concerned field of the researchers. In conclusion, the number of publications on myocardial reperfusion injury increases during the past three decades, which mainly focused on prognosis, mechanism, and treatment. Prognosis is the most concerned field, whereas studies on mechanism and treatment are relatively lacking.


2018 ◽  
Vol 14 (4) ◽  
pp. 734-747 ◽  
Author(s):  
Constance de Saint Laurent

There has been much hype, over the past few years, about the recent progress of artificial intelligence (AI), especially through machine learning. If one is to believe many of the headlines that have proliferated in the media, as well as in an increasing number of scientific publications, it would seem that AI is now capable of creating and learning in ways that are starting to resemble what humans can do. And so that we should start to hope – or fear – that the creation of fully cognisant machine might be something we will witness in our life time. However, much of these beliefs are based on deep misconceptions about what AI can do, and how. In this paper, I start with a brief introduction to the principles of AI, machine learning, and neural networks, primarily intended for psychologists and social scientists, who often have much to contribute to the debates surrounding AI but lack a clear understanding of what it can currently do and how it works. I then debunk four common myths associated with AI: 1) it can create, 2) it can learn, 3) it is neutral and objective, and 4) it can solve ethically and/or culturally sensitive problems. In a third and last section, I argue that these misconceptions represent four main dangers: 1) avoiding debate, 2) naturalising our biases, 3) deresponsibilising creators and users, and 4) missing out some of the potential uses of machine learning. I finally conclude on the potential benefits of using machine learning in research, and thus on the need to defend machine learning without romanticising what it can actually do.


2019 ◽  
Vol 6 (1) ◽  
Author(s):  
Claus Boye Asmussen ◽  
Charles Møller

Abstract Manual exploratory literature reviews should be a thing of the past, as technology and development of machine learning methods have matured. The learning curve for using machine learning methods is rapidly declining, enabling new possibilities for all researchers. A framework is presented on how to use topic modelling on a large collection of papers for an exploratory literature review and how that can be used for a full literature review. The aim of the paper is to enable the use of topic modelling for researchers by presenting a step-by-step framework on a case and sharing a code template. The framework consists of three steps; pre-processing, topic modelling, and post-processing, where the topic model Latent Dirichlet Allocation is used. The framework enables huge amounts of papers to be reviewed in a transparent, reliable, faster, and reproducible way.


2021 ◽  
pp. 297-315
Author(s):  
Alireza Tamaddoni-Nezhad ◽  
David Bohan ◽  
Ghazal Afroozi Milani ◽  
Alan Raybould ◽  
Stephen Muggleton

Humanity is facing existential, societal challenges related to food security, ecosystem conservation, antimicrobial resistance, etc, and Artificial Intelligence (AI) is already playing an important role in tackling these new challenges. Most current AI approaches are limited when it comes to ‘knowledge transfer’ with humans, i.e. it is difficult to incorporate existing human knowledge and also the output knowledge is not human comprehensible. In this chapter we demonstrate how a combination of comprehensible machine learning, text-mining and domain knowledge could enhance human-machine collaboration for the purpose of automated scientific discovery where humans and computers jointly develop and evaluate scientific theories. As a case study, we describe a combination of logic-based machine learning (which included human-encoded ecological background knowledge) and text-mining from scientific publications (to verify machine-learned hypotheses) for the purpose of automated discovery of ecological interaction networks (food-webs) to detect change in agricultural ecosystems using the Farm Scale Evaluations (FSEs) of genetically modified herbicide-tolerant (GMHT) crops dataset. The results included novel food-web hypotheses, some confirmed by subsequent experimental studies (e.g. DNA analysis) and published in scientific journals. These machine-leaned food-webs were also used as the basis of a recent study revealing resilience of agro-ecosystems to changes in farming management using GMHT crops.


Author(s):  
Aleksey Klokov ◽  
Evgenii Slobodyuk ◽  
Michael Charnine

The object of the research when writing the work was the body of text data collected together with the scientific advisor and the algorithms for processing the natural language of analysis. The stream of hypotheses has been tested against computer science scientific publications through a series of simulation experiments described in this dissertation. The subject of the research is algorithms and the results of the algorithms, aimed at predicting promising topics and terms that appear in the course of time in the scientific environment. The result of this work is a set of machine learning models, with the help of which experiments were carried out to identify promising terms and semantic relationships in the text corpus. The resulting models can be used for semantic processing and analysis of other subject areas.


2019 ◽  
Vol 19 (1) ◽  
pp. 4-16 ◽  
Author(s):  
Qihui Wu ◽  
Hanzhong Ke ◽  
Dongli Li ◽  
Qi Wang ◽  
Jiansong Fang ◽  
...  

Over the past decades, peptide as a therapeutic candidate has received increasing attention in drug discovery, especially for antimicrobial peptides (AMPs), anticancer peptides (ACPs) and antiinflammatory peptides (AIPs). It is considered that the peptides can regulate various complex diseases which are previously untouchable. In recent years, the critical problem of antimicrobial resistance drives the pharmaceutical industry to look for new therapeutic agents. Compared to organic small drugs, peptide- based therapy exhibits high specificity and minimal toxicity. Thus, peptides are widely recruited in the design and discovery of new potent drugs. Currently, large-scale screening of peptide activity with traditional approaches is costly, time-consuming and labor-intensive. Hence, in silico methods, mainly machine learning approaches, for their accuracy and effectiveness, have been introduced to predict the peptide activity. In this review, we document the recent progress in machine learning-based prediction of peptides which will be of great benefit to the discovery of potential active AMPs, ACPs and AIPs.


Entropy ◽  
2021 ◽  
Vol 23 (3) ◽  
pp. 300
Author(s):  
Mark Lokanan ◽  
Susan Liu

Protecting financial consumers from investment fraud has been a recurring problem in Canada. The purpose of this paper is to predict the demographic characteristics of investors who are likely to be victims of investment fraud. Data for this paper came from the Investment Industry Regulatory Organization of Canada’s (IIROC) database between January of 2009 and December of 2019. In total, 4575 investors were coded as victims of investment fraud. The study employed a machine-learning algorithm to predict the probability of fraud victimization. The machine learning model deployed in this paper predicted the typical demographic profile of fraud victims as investors who classify as female, have poor financial knowledge, know the advisor from the past, and are retired. Investors who are characterized as having limited financial literacy but a long-time relationship with their advisor have reduced probabilities of being victimized. However, male investors with low or moderate-level investment knowledge were more likely to be preyed upon by their investment advisors. While not statistically significant, older adults, in general, are at greater risk of being victimized. The findings from this paper can be used by Canadian self-regulatory organizations and securities commissions to inform their investors’ protection mandates.


Entropy ◽  
2021 ◽  
Vol 23 (5) ◽  
pp. 626
Author(s):  
Ramya Gupta ◽  
Abhishek Prasad ◽  
Suresh Babu ◽  
Gitanjali Yadav

A global event such as the COVID-19 crisis presents new, often unexpected responses that are fascinating to investigate from both scientific and social standpoints. Despite several documented similarities, the coronavirus pandemic is clearly distinct from the 1918 flu pandemic in terms of our exponentially increased, almost instantaneous ability to access/share information, offering an unprecedented opportunity to visualise rippling effects of global events across space and time. Personal devices provide “big data” on people’s movement, the environment and economic trends, while access to the unprecedented flurry in scientific publications and media posts provides a measure of the response of the educated world to the crisis. Most bibliometric (co-authorship, co-citation, or bibliographic coupling) analyses ignore the time dimension, but COVID-19 has made it possible to perform a detailed temporal investigation into the pandemic. Here, we report a comprehensive network analysis based on more than 20,000 published documents on viral epidemics, authored by over 75,000 individuals from 140 nations in the past one year of the crisis. Unlike the 1918 flu pandemic, access to published data over the past two decades enabled a comparison of publishing trends between the ongoing COVID-19 pandemic and those of the 2003 SARS epidemic to study changes in thematic foci and societal pressures dictating research over the course of a crisis.


Sign in / Sign up

Export Citation Format

Share Document