scholarly journals Innovative Empirical Model for Predicting National Banks’ Financial Failure with Artificial Intelligence Subset Data Analysis in the United States

2020 ◽  
Vol 3 (1) ◽  
pp. 98-111
Author(s):  
Karina Kasztelnik

AbstractThe principal objective of this research study was to investigate the impact of the Great Economic Recession of 2008 on national banks’ equity investment valuations and create an empirical model for predicting national banks’ financial failure in the United States. The focal period of the study was from 2009 to 2012, and public data sources used. It is not known to what extent national banks’ stock value investments are based on the return on equity. This causal-comparative study explores the degree to which national banks’ value investment in terms of the price to earnings ratio impacts their return on equity and the extent to which these banks’ stock value investment in terms of dividend yield impacts their return on equity. We used statistical modeling and the machine learning model to find hidden patterns in the input data. The principal finding of this research is that the median earnings per share in 2012 and the dividend yield in 2009 were significantly larger than the median return on equity in 2009 and 2012. Additionally, the dividend yield in 2012 was significantly smaller than the median return on equity in 2012. These findings can contribute to improving our understanding of how banks can predict financial failure using the new machine learning features of artificial intelligence to build an early warning system with the innovative risk measurement tool.

2021 ◽  
Author(s):  
satya katragadda ◽  
ravi teja bhupatiraju ◽  
vijay raghavan ◽  
ziad ashkar ◽  
raju gottumukkala

Abstract Background: Travel patterns of humans play a major part in the spread of infectious diseases. This was evident in the geographical spread of COVID-19 in the United States. However, the impact of this mobility and the transmission of the virus due to local travel, compared to the population traveling across state boundaries, is unknown. This study evaluates the impact of local vs. visitor mobility in understanding the growth in the number of cases for infectious disease outbreaks. Methods: We use two different mobility metrics, namely the local risk and visitor risk extracted from trip data generated from anonymized mobile phone data across all 50 states in the United States. We analyzed the impact of just using local trips on infection spread and infection risk potential generated from visitors' trips from various other states. We used the Diebold-Mariano test to compare across three machine learning models. Finally, we compared the performance of models, including visitor mobility for all the three waves in the United States and across all 50 states. Results: We observe that visitor mobility impacts case growth and that including visitor mobility in forecasting the number of COVID-19 cases improves prediction accuracy by 34. We found the statistical significance with respect to the performance improvement resulting from including visitor mobility using the Diebold-Mariano test. We also observe that the significance was much higher during the first peak March to June 2020. Conclusion: With presence of cases everywhere (i.e. local and visitor), visitor mobility (even within the country) is shown to have significant impact on growth in number of cases. While it is not possible to account for other factors such as the impact of interventions, and differences in local mobility and visitor mobility, we find that these observations can be used to plan for both reopening and limiting visitors from regions where there are high number of cases.


2020 ◽  
pp. 97-102
Author(s):  
Benjamin Wiggins

Can risk assessment be made fair? The conclusion of Calculating Race returns to actuarial science’s foundations in probability. The roots of probability rest in a pair of problems posed to Blaise Pascal and Pierre de Fermat in the summer of 1654: “the Dice Problem” and “the Division Problem.” From their very foundation, the mathematics of probability offered the potential not only to be used to gain an advantage (as in the case of the Dice Problem), but also to divide material fairly (as in the case of the Division Problem). As the United States and the world enter an age driven by Big Data, algorithms, artificial intelligence, and machine learning and characterized by an actuarialization of everything, we must remember that risk assessment need not be put to use for individual, corporate, or government advantage but, rather, that it has always been capable of guiding how to distribute risk equitably instead.


10.2196/31983 ◽  
2021 ◽  
Vol 1 (1) ◽  
pp. e31983
Author(s):  
Arriel Benis ◽  
Anat Chatsubi ◽  
Eugene Levner ◽  
Shai Ashkenazi

Background Discussions of health issues on social media are a crucial information source reflecting real-world responses regarding events and opinions. They are often important in public health care, since these are influencing pathways that affect vaccination decision-making by hesitant individuals. Artificial intelligence methodologies based on internet search engine queries have been suggested to detect disease outbreaks and population behavior. Among social media, Twitter is a common platform of choice to search and share opinions and (mis)information about health care issues, including vaccination and vaccines. Objective Our primary objective was to support the design and implementation of future eHealth strategies and interventions on social media to increase the quality of targeted communication campaigns and therefore increase influenza vaccination rates. Our goal was to define an artificial intelligence–based approach to elucidate how threads in Twitter on influenza vaccination changed during the COVID-19 pandemic. Such findings may support adapted vaccination campaigns and could be generalized to other health-related mass communications. Methods The study comprised the following 5 stages: (1) collecting tweets from Twitter related to influenza, vaccines, and vaccination in the United States; (2) data cleansing and storage using machine learning techniques; (3) identifying terms, hashtags, and topics related to influenza, vaccines, and vaccination; (4) building a dynamic folksonomy of the previously defined vocabulary (terms and topics) to support the understanding of its trends; and (5) labeling and evaluating the folksonomy. Results We collected and analyzed 2,782,720 tweets of 420,617 unique users between December 30, 2019, and April 30, 2021. These tweets were in English, were from the United States, and included at least one of the following terms: “flu,” “influenza,” “vaccination,” “vaccine,” and “vaxx.” We noticed that the prevalence of the terms vaccine and vaccination increased over 2020, and that “flu” and “covid” occurrences were inversely correlated as “flu” disappeared over time from the tweets. By combining word embedding and clustering, we then identified a folksonomy built around the following 3 topics dominating the content of the collected tweets: “health and medicine (biological and clinical aspects),” “protection and responsibility,” and “politics.” By analyzing terms frequently appearing together, we noticed that the tweets were related mainly to COVID-19 pandemic events. Conclusions This study focused initially on vaccination against influenza and moved to vaccination against COVID-19. Infoveillance supported by machine learning on Twitter and other social media about topics related to vaccines and vaccination against communicable diseases and their trends can lead to the design of personalized messages encouraging targeted subpopulations’ engagement in vaccination. A greater likelihood that a targeted population receives a personalized message is associated with higher response, engagement, and proactiveness of the target population for the vaccination process.


Author(s):  
Debra A. Murphy ◽  
William D. Marelich ◽  
Dannie Hoffman ◽  
Mark A. Schuster

2020 ◽  
Author(s):  
Xiaoqian Jiang ◽  
Lishan Yu ◽  
Hamisu M. Salihub ◽  
Deepa Dongarwar

BACKGROUND In the United States, State laws require birth certificates to be completed for all births; and federal law mandates national collection and publication of births and other vital statistics data. National Center for Health Statistics (NCHS) has published the key statistics of birth data over the years. These data files, from as early as the 1970s, have been released and made publicly available. There are about 3 million new births each year, and every birth is a record in the data set described by hundreds of variables. The total data cover more than half of the current US population, making it an invaluable resource to study and examine birth epidemiology. Using such big data, researchers can ask interesting questions and study longitudinal patterns, for example, the impact of mother's drinking status to infertility in metropolitans in the last decade, or the education level of the biological father to the c-sections over the years. However, existing published data sets cannot directly support these research questions as there are adjustments to the variables and their categories, which makes these individually published data files fragmented. The information contained in the published data files is highly diverse, containing hundreds of variables each year. Besides minor adjustments like renaming and increasing variable categories, some major updates significantly changed the fields of statistics (including removal, addition, and modification of the variables), making the published data disconnected and ambiguous to use over multiple years. Researchers have previously reconstructed features to study temporal patterns, but the scale is limited (focusing only on a few variables of interest). Many have reinvented the wheels, and such reconstructions lack consistency as different researchers might use different criteria to harmonize variables, leading to inconsistent findings and limiting the reproducibility of research. There is no systematic effort to combine about five decades of data files into a database that includes every variable that has ever been released by NCHS. OBJECTIVE To utilize machine learning techniques to combine the United States (US) natality data for the last five decades, with changing variables and factors, into a consistent database. METHODS We developed a feasible and efficient deep-learning-based framework to harmonize data sets of live births in the US from 1970 to 2018. We constructed a graph based on the property and elements of databases including variables and conducted a graph convolutional network (GCN) on the graph to learn the graph embeddings for nodes where the learned embeddings implied the similarity of variables. We devised a novel loss function with a slack margin and a banlist mechanism (for a random walk) to learn the desired structure (two nodes sharing more information were more similar to each other.). We developed an active learning mechanism to conduct the harmonization. RESULTS We harmonized historical US birth data and resolved conflicts in ambiguous terms. From a total of 9,321 variables (i.e., 783 stemmed variables, from 1970 to 2018) we applied our model iteratively together with human review, obtaining 323 hyperchains of variables. Hyperchains for harmonization were composed of 201 stemmed variable pairs when considering any pairs of different stemmed variables changed over years. During the harmonization, the first round of our model provided 305 candidates stemmed variable pairs (based on the top-20 most similar variables of each variable based on the learned embeddings of variables) and achieved recall and precision of 87.56%, 57.70%, respectively. CONCLUSIONS Our harmonized graph neural network (HGNN) method provides a feasible and efficient way to connect relevant databases at a meta-level. Adapting to databases' property and characteristics, HGNN can learn patterns and search relations globally, which is powerful to discover the similarity between variables among databases. Smart utilization of machine learning can significantly reduce the manual effort in database harmonization and integration of fragmented data into useful databases for future research.


2021 ◽  
Author(s):  
Arriel Benis ◽  
Anat Chatsubi ◽  
Eugene Levner ◽  
Shai Ashkenazi

BACKGROUND Discussions of health issues on social media are a crucial information source reflecting real-world responses regarding events and opinions. They are often important in public health care, since these are influencing pathways that affect vaccination decision-making by hesitant individuals. Artificial intelligence methodologies based on internet search engine queries have been suggested to detect disease outbreaks and population behavior. Among social media, Twitter is a common platform of choice to search and share opinions and (mis)information about health care issues, including vaccination and vaccines. OBJECTIVE Our primary objective was to support the design and implementation of future eHealth strategies and interventions on social media to increase the quality of targeted communication campaigns and therefore increase influenza vaccination rates. Our goal was to define an artificial intelligence–based approach to elucidate how threads in Twitter on influenza vaccination changed during the COVID-19 pandemic. Such findings may support adapted vaccination campaigns and could be generalized to other health-related mass communications. METHODS The study comprised the following 5 stages: (1) collecting tweets from Twitter related to influenza, vaccines, and vaccination in the United States; (2) data cleansing and storage using machine learning techniques; (3) identifying terms, hashtags, and topics related to influenza, vaccines, and vaccination; (4) building a dynamic folksonomy of the previously defined vocabulary (terms and topics) to support the understanding of its trends; and (5) labeling and evaluating the folksonomy. RESULTS We collected and analyzed 2,782,720 tweets of 420,617 unique users between December 30, 2019, and April 30, 2021. These tweets were in English, were from the United States, and included at least one of the following terms: “flu,” “influenza,” “vaccination,” “vaccine,” and “vaxx.” We noticed that the prevalence of the terms vaccine and vaccination increased over 2020, and that “flu” and “covid” occurrences were inversely correlated as “flu” disappeared over time from the tweets. By combining word embedding and clustering, we then identified a folksonomy built around the following 3 topics dominating the content of the collected tweets: “health and medicine (biological and clinical aspects),” “protection and responsibility,” and “politics.” By analyzing terms frequently appearing together, we noticed that the tweets were related mainly to COVID-19 pandemic events. CONCLUSIONS This study focused initially on vaccination against influenza and moved to vaccination against COVID-19. Infoveillance supported by machine learning on Twitter and other social media about topics related to vaccines and vaccination against communicable diseases and their trends can lead to the design of personalized messages encouraging targeted subpopulations’ engagement in vaccination. A greater likelihood that a targeted population receives a personalized message is associated with higher response, engagement, and proactiveness of the target population for the vaccination process.


2020 ◽  
Vol 21 (1) ◽  
pp. 158-177
Author(s):  
Rajesh Chakrabarti ◽  
Kaushiki Sanyal

The impact of artificial intelligence (AI) on every aspect of our lives is inevitable and already being felt in numerous ways. Countries are grappling with the opportunities and challenges that AI presents. Among the South Asian countries, India has taken a lead in promoting and regulating AI. However, it lags significantly behind countries such as China or the United States. This article explores India’s AI ecosystem, the threats and challenges it faces, and the ethical issues it needs to consider. Finally, it examines the common concerns among South Asian nations and the possibility of coming together to promote and regulate AI in the region. JEL: Z: Z0: Z000


Sign in / Sign up

Export Citation Format

Share Document