Weka meets TraceLab: Toward convenient classification: Machine learning for requirements engineering problems: A position paper

Author(s):  
Jane Huffman Hayes ◽  
Wenbin Li ◽  
Mona Rahimi
Fuels ◽  
2021 ◽  
Vol 2 (3) ◽  
pp. 286-303
Author(s):  
Vuong Van Pham ◽  
Ebrahim Fathi ◽  
Fatemeh Belyadi

The success of machine learning (ML) techniques implemented in different industries heavily rely on operator expertise and domain knowledge, which is used in manually choosing an algorithm and setting up the specific algorithm parameters for a problem. Due to the manual nature of model selection and parameter tuning, it is impossible to quantify or evaluate the quality of this manual process, which in turn limits the ability to perform comparison studies between different algorithms. In this study, we propose a new hybrid approach for developing machine learning workflows to help automated algorithm selection and hyperparameter optimization. The proposed approach provides a robust, reproducible, and unbiased workflow that can be quantified and validated using different scoring metrics. We have used the most common workflows implemented in the application of artificial intelligence (AI) and ML in engineering problems including grid/random search, Bayesian search and optimization, genetic programming, and compared that with our new hybrid approach that includes the integration of Tree-based Pipeline Optimization Tool (TPOT) and Bayesian optimization. The performance of each workflow is quantified using different scoring metrics such as Pearson correlation (i.e., R2 correlation) and Mean Square Error (i.e., MSE). For this purpose, actual field data obtained from 1567 gas wells in Marcellus Shale, with 121 features from reservoir, drilling, completion, stimulation, and operation is tested using different proposed workflows. A proposed new hybrid workflow is then used to evaluate the type well used for evaluation of Marcellus shale gas production. In conclusion, our automated hybrid approach showed significant improvement in comparison to other proposed workflows using both scoring matrices. The new hybrid approach provides a practical tool that supports the automated model and hyperparameter selection, which is tested using real field data that can be implemented in solving different engineering problems using artificial intelligence and machine learning. The new hybrid model is tested in a real field and compared with conventional type wells developed by field engineers. It is found that the type well of the field is very close to P50 predictions of the field, which shows great success in the completion design of the field performed by field engineers. It also shows that the field average production could have been improved by 8% if shorter cluster spacing and higher proppant loading per cluster were used during the frac jobs.


Energies ◽  
2021 ◽  
Vol 14 (4) ◽  
pp. 930
Author(s):  
Fahimeh Hadavimoghaddam ◽  
Mehdi Ostadhassan ◽  
Ehsan Heidaryan ◽  
Mohammad Ali Sadri ◽  
Inna Chapanova ◽  
...  

Dead oil viscosity is a critical parameter to solve numerous reservoir engineering problems and one of the most unreliable properties to predict with classical black oil correlations. Determination of dead oil viscosity by experiments is expensive and time-consuming, which means developing an accurate and quick prediction model is required. This paper implements six machine learning models: random forest (RF), lightgbm, XGBoost, multilayer perceptron (MLP) neural network, stochastic real-valued (SRV) and SuperLearner to predict dead oil viscosity. More than 2000 pressure–volume–temperature (PVT) data were used for developing and testing these models. A huge range of viscosity data were used, from light intermediate to heavy oil. In this study, we give insight into the performance of different functional forms that have been used in the literature to formulate dead oil viscosity. The results show that the functional form f(γAPI,T), has the best performance, and additional correlating parameters might be unnecessary. Furthermore, SuperLearner outperformed other machine learning (ML) algorithms as well as common correlations that are based on the metric analysis. The SuperLearner model can potentially replace the empirical models for viscosity predictions on a wide range of viscosities (any oil type). Ultimately, the proposed model is capable of simulating the true physical trend of the dead oil viscosity with variations of oil API gravity, temperature and shear rate.


2021 ◽  
Vol 22 (3) ◽  
pp. 313-320
Author(s):  
Dana Petcu

This position paper aims to identify the current and future challenges in application, workload or service deployment mechanisms in Cloud-to-Edge environments. We argue that the adoption of the microservices and unikernels on large scale is adding new entries on the list of requirements of a deployment mechanism, but offers an opportunity to decentralize the associated processes and improve the scalability of the applications. Moreover, the deployment in Cloud-to-Edge environment needs the support of federated machine learning.


2014 ◽  
Vol 721 ◽  
pp. 750-753
Author(s):  
Jian Sheng Pan ◽  
Shi Cheng

Statisticians agree that signed epistemologies are an interesting new topic in the field of machine learning, and cyberneticists concur. Given the current status of pseudorandom configurations, cryptographers famously desire the refinement of simulated annealing. Our focus in this position paper is not on whether superblocks and extreme programming can collaborate to answer this quagmire, but rather on introducing a methodology for modular information (Timer).


Author(s):  
Yoram Reich

Since the inception of research on machine learning (ML), these techniques have been associated with the task of automated knowledge generation or knowledge reorganization. This association still prevails, as seen in this issue. When the use of ML programs began to attract researchers in engineering design, different existing tools were used to test their utility and gradually, variations of these tools and methods have sprung up. In many cases, the use of these tools was based on availability and not necessarily applicability. When we began working on ML in design, we attempted to follow a different path (Reich, 1991a; Reich & Fenves, 1992) that led to the design of Bridger (Reich & Fenves, 1995), a system for learning bridge synthesis knowledge. Subsequent experiences and further reflection led us to conclude that the process of using ML in design requires careful and systematic treatment for identifying appropriate ML programs for executing the learning tasks we wish to perform (Reich, 1991b, 1993a). Another observation was that the task of creating or reorganizing knowledge for real design tasks is outside the scope of present ML programs. Establishing the practical importance of ML techniques had to start by addressing engineering problems that could benefit from present ML programs.


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Amirhessam Tahmassebi ◽  
Mehrtash Motamedi ◽  
Amir H. Alavi ◽  
Amir H. Gandomi

PurposeEngineering design and operational decisions depend largely on deep understanding of applications that requires assumptions for simplification of the problems in order to find proper solutions. Cutting-edge machine learning algorithms can be used as one of the emerging tools to simplify this process. In this paper, we propose a novel scalable and interpretable machine learning framework to automate this process and fill the current gap.Design/methodology/approachThe essential principles of the proposed pipeline are mainly (1) scalability, (2) interpretibility and (3) robust probabilistic performance across engineering problems. The lack of interpretibility of complex machine learning models prevents their use in various problems including engineering computation assessments. Many consumers of machine learning models would not trust the results if they cannot understand the method. Thus, the SHapley Additive exPlanations (SHAP) approach is employed to interpret the developed machine learning models.FindingsThe proposed framework can be applied to a variety of engineering problems including seismic damage assessment of structures. The performance of the proposed framework is investigated using two case studies of failure identification in reinforcement concrete (RC) columns and shear walls. In addition, the reproducibility, reliability and generalizability of the results were validated and the results of the framework were compared to the benchmark studies. The results of the proposed framework outperformed the benchmark results with high statistical significance.Originality/valueAlthough, the current study reveals that the geometric input features and reinforcement indices are the most important variables in failure modes detection, better model can be achieved with employing more robust strategies to establish proper database to decrease the errors in some of the failure modes identification.


Sign in / Sign up

Export Citation Format

Share Document