learning from data
Recently Published Documents


TOTAL DOCUMENTS

162
(FIVE YEARS 55)

H-INDEX

18
(FIVE YEARS 4)

Electronics ◽  
2022 ◽  
Vol 11 (2) ◽  
pp. 189
Author(s):  
Álvaro de Pablo ◽  
Oscar Araque ◽  
Carlos A. Iglesias

The analysis of the content of posts written on social media has established an important line of research in recent years. The study of these texts, as well as their relationship with each other and their dependence on the platform on which they are written, enables the behavior analysis of users and their opinions with respect to different domains. In this work, a hybrid machine learning-based system has been developed to classify texts using topic modeling techniques and different word-vector representations, as well as traditional text representations. The system has been trained with ride-hailing posts extracted from Reddit, showing promising performance. Then, the generated models have been tested with data extracted from other sources such as Twitter and Google Play, classifying these texts without retraining any models and thus performing Transfer Learning. The obtained results show that our proposed architecture is effective when performing Transfer Learning from data-rich domains and applying them to other sources.


2021 ◽  
Vol 12 (1) ◽  
Author(s):  
Shih-Han Wang ◽  
Hemanth Somarajan Pillai ◽  
Siwen Wang ◽  
Luke E. K. Achenie ◽  
Hongliang Xin

AbstractDespite recent advances of data acquisition and algorithms development, machine learning (ML) faces tremendous challenges to being adopted in practical catalyst design, largely due to its limited generalizability and poor explainability. Herein, we develop a theory-infused neural network (TinNet) approach that integrates deep learning algorithms with the well-established d-band theory of chemisorption for reactivity prediction of transition-metal surfaces. With simple adsorbates (e.g., *OH, *O, and *N) at active site ensembles as representative descriptor species, we demonstrate that the TinNet is on par with purely data-driven ML methods in prediction performance while being inherently interpretable. Incorporation of scientific knowledge of physical interactions into learning from data sheds further light on the nature of chemical bonding and opens up new avenues for ML discovery of novel motifs with desired catalytic properties.


2021 ◽  
Vol 28 (1) ◽  
pp. e100447
Author(s):  
Davy van de Sande ◽  
Michel E. Van Genderen ◽  
Joost Huiskens ◽  
Robert E. R. Veen ◽  
Yvonne Meijerink ◽  
...  

Introduction In the current situation, clinical patient data are often siloed in multiple hospital information systems. Especially in the intensive care unit (ICU), large volumes of clinical data are routinely collected through continuous patient monitoring. Although these data often contain useful information for clinical decision making, they are not frequently used to improve quality of care. During, but also after, pressing times, data-driven methods can be used to mine treatment patterns from clinical data to determine the best treatment options from a hospitals own clinical data.Methods In this implementer report, we describe how we implemented a data infrastructure that enabled us to learn in real time from consecutive COVID-19 ICU admissions. In addition, we explain our step-by-step multidisciplinary approach to establish such a data infrastructure.Conclusion By sharing our steps and approach, we aim to inspire others, in and outside ICU walls, to make more efficient use of data at hand, now and in the future.


2021 ◽  
Author(s):  
Mario Ramos-Maldonado ◽  
Cristhian Aguilera-Carrasco

Wood industry is key for sustainability and an important economic activity in many countries. In manufacturing plants, wood variability turns operation management more complex. In a competitive scenario, assets availability is critical to achieve higher productivity. In a new fourth industrial revolution, Industry 4.0, data engineering permits efficient decisions making. Phenomena difficult to model with conventional techniques are turned possible with algorithms based on artificial intelligence. Sensors and machine learning techniques allow intelligent analysis of data. However, algorithms are highly sensitive of the problem and his study to decide on which work is critical. For the manufacturing wood processes, Industry 4.0 is a great opportunity. Wood is a material of biological origin and generates variabilities over the manufacturing processes. For example, in the veneer drying, density and anatomical structure impact the product quality. Scanners have been developed to measure variables and outcomes, but decisions are made yet by humans. Today, robust sensors, computing capacity, communications and intelligent algorithms permit to manage wood variability. Real-time actions can be achieved by learning from data. This paper presents trends and opportunities provided by Industry 4.0 components. Sensors, decision support systems and intelligent algorithms use are reviewed. Some applications are presented.


2021 ◽  
Vol 68 (1) ◽  
pp. 1-16
Author(s):  
Józef Pociecha

The starting point for the presentation of the similarities and differences between the principles of conducting statistical research according to the rules of both statistical inference and statistical learning is the paradigm theory, formulated by Thomas Kuhn. In the first section of this paper, the essential features of the statistical inference paradigm are characterised, with particular attention devoted to its limitations in contemporary statistical research. Subsequently, the article presents the challenges faced by this research jointly with the expanding opportunities for their effective reduction. The essence of learning from data is discussed and the principles of statistical learning are defined. Moreover, significant features of the statistical learning paradigm are formulated in the context of the differences between the statistical inference paradigm and the statistical learning paradigm. It is emphasised that the statistical learning paradigm, as the more universal one of the two discussed, broadens the possibilities of conducting statistical research, especially in socio-economic sciences.


2021 ◽  
Vol 26 (5) ◽  
pp. 1-25
Author(s):  
Chin-Hsien Wu ◽  
Hao-Wei Zhang ◽  
Chia-Wei Liu ◽  
Ta-Ching Yu ◽  
Chi-Yen Yang

With the progress of the manufacturing process, NAND flash memory has evolved from the single-level cell and multi-level cell into the triple-level cell (TLC). NAND flash memory has physical problems such as the characteristic of erase-before-write and the limitation of program/erase cycles. Moreover, TLC NAND flash memory has low reliability and short lifetime. Thus, we propose a dynamic Huffman coding method that can apply to the write operations of NAND flash memory. The proposed method exploits observations from a Huffman tree and machine learning from data patterns to dynamically select a suitable Huffman coding. According to the experimental results, the proposed method can improve the reliability of TLC NAND flash memory and also consider the compression performance for those applications that require the Huffman coding.


2021 ◽  
Author(s):  
Stephen Gilbert ◽  
Matthew Fenech ◽  
Martin Hirsch ◽  
Shubhanan Upadhyay ◽  
Andrea Biasiucci ◽  
...  

UNSTRUCTURED One of the greatest strengths of artificial intelligence and machine learning (AI/ML) approaches in healthcare is that their performance can be continually improved based on updates from automated learning from data. However, healthcare AI/ML models are currently essentially regulated under provisions that were developed for an earlier age of slowly updated medical devices - requiring major documentation reshape and re-validation with every major update of the model generated by the ML algorithm. This creates minor problems for models that will be re-trained and updated only occasionally, but major problems for models that will learn from data in real-time or near real-time. Regulators have announced action plans for fundamental changes in regulatory approaches. Here, we examine the current regulatory frameworks and the developments in this domain. The status quo and recent developments are reviewed, and we argue that these innovative approaches to healthcare need these matching innovative approaches to regulation and that these approaches will bring benefits for patients. International perspectives from the WHO, and the FDA’s proposed approach, based around oversight of tool developers’ quality management systems and defined algorithm change protocols, offer a much-needed paradigm shift, and strive for a balanced approach to enabling rapid improvements in healthcare through AI innovation, whilst simultaneously ensuring patient safety. The draft EU regulatory framework indicates similar approaches, but no detail has yet been provided on how algorithm change protocols will be implemented in the EU, and this is required for the full benefits of AI/ML-based innovation for patients and for EU healthcare systems to be realised.


2021 ◽  
Vol 11 (7) ◽  
pp. 3045-3077
Author(s):  
Cristina C. B. Cavalcante ◽  
Cid C. de Souza ◽  
Célio Maschio ◽  
Denis Schiozer ◽  
Anderson Rocha

AbstractHistory matching is an important reservoir engineering process whereby the values of uncertain attributes of a reservoir model are changed to find models that have a better chance of reproducing the performance of an actual reservoir. As a typical inverse and ill-posed problem, different combinations of reservoir uncertain attributes lead to equally well-matched models and the success of a history-matching approach is usually measured in terms of its ability to efficiently find multiple history-matched models inside the search space defined by the parameterization of the problem (multiple-matched models have a higher chance of better representing the reservoir performance forecast). While studies on history-matching approaches have produced remarkable progress over the last two decades, given the uniqueness of each reservoir’s history-matching problem, no strategy is proven effective for all cases, and finding alternative, efficient, and effective history-matching methodologies is still a research challenge. In this work, we introduce a learning-from-data approach with path relinking and soft clustering to the history-matching problem. The proposed algorithm is designed to learn the patterns of input attributes that are associated with good matching quality from the set of available solutions, and has two stages that handle different types of reservoir uncertain attributes. In each stage, the algorithm evaluates the data of all-available solutions continuously and, based on the acquired information, dynamically decides what needs to be changed, where the changes shall take place, and how such changes will occur in order to generate new (and hopefully better) solutions. We validate our approach using the UNISIM-I-H benchmark, a complex synthetic case constructed with real data from the Namorado Field, Campos Basin, Brazil. Experimental results indicate the potential of the proposed approach in finding models with significantly better history-matching quality. Considering a global misfit quality metric, the final best solutions found by our approach are up to 77% better than the corresponding initial best solutions in the datasets used in the experiments. Moreover, compared with previous work for the same benchmark, the proposed learning-from-data approach is competitive regarding the quality of solutions found and, above all, it offers a significant reduction (up to 30 × less) in the number of simulations.


Sign in / Sign up

Export Citation Format

Share Document