Intelligent algorithms are recently used in the optimization process in chemical engineering and application of multiphase flows such as bubbling flow. This overview of modeling can be a great replacement with complex numerical methods or very time-consuming and disruptive measurement experimental process. In this study, we develop the adaptive network-based fuzzy inference system (ANFIS) method for mapping inputs and outputs together and understand the behavior of the fluid flow from other output parameters of the bubble column reactor. Neural cells can fully learn the process in their memory and after the training stage, the fuzzy structure predicts the multiphase flow data. Four inputs such as x coordinate, y coordinate, z coordinate, and air superficial velocity and one output such as pressure gradient are considered in the learning process of the ANFIS method. During the learning process, the different number of the membership function, type of membership functions and the number of inputs are examined to achieve the intelligent algorithm with high accuracy. The results show that as the number of inputs increases the accuracy of the ANFIS method rises up to R^2>0.99 almost for all cases, while the increment in the number of rules has a effect on the intelligence of artificial algorithm. This finding shows that the density of neural objects or higher input parameters enables the moded for better understanding. We also proposed a new evaluation of data in the bubble column reactor by mapping inputs and outputs and shuffle all parameters together to understand the behaviour of the multiphase flow as a function of either inputs or outputs. This new process of mapping inputs and outputs data provides a framework to fully understand the flow in the fluid domain in a short time of fuzzy structure calculation.
Machine Learning (ML) is one of the most exciting and dynamic areas of modern research and application. The purpose of this review is to provide an introduction to the core concepts and tools of machine learning in a manner easily understood and intuitive to physicists. The review begins by covering fundamental concepts in ML and modern statistics such as the bias-variance tradeoff, overfitting, regularization, and generalization before moving on to more advanced topics in both supervised and unsupervised learning. Topics covered in the review include ensemble models, deep learning and neural networks, clustering and data visualization, energy-based models (including MaxEnt models and Restricted Boltzmann Machines), and variational methods. Throughout, we emphasize the many natural connections between ML and statistical physics. A notable aspect of the review is the use of Python notebooks to introduce modern ML/statistical packages to readers using physics-inspired datasets (the Ising Model and Monte-Carlo simulations of supersymmetric decays of proton-proton collisions). We conclude with an extended outlook discussing possible uses of machine learning for furthering our understanding of the physical world as well as open problems in ML where physicists maybe able to contribute. (Notebooks are available at https://physics.bu.edu/~pankajm/MLnotebooks.html )
Recent advances in graph convolutional networks have significantly improved the performance of chemical predictions, raising a new research question: "how do we explain the predictions of graph convolutional networks?" A possible approach to answer this question is to visualize evidence substructures responsible for the predictions. For chemical property prediction tasks, the sample size of the training data is often small and/or a label imbalance problem occurs, where a few samples belong to a single class and the majority of samples belong to the other classes. This can lead to uncertainty related to the learned parameters of the machine learning model. To address this uncertainty, we propose BayesGrad, utilizing the Bayesian predictive distribution, to define the importance of each node in an input graph, which is computed efficiently using the dropout technique. We demonstrate that BayesGrad successfully visualizes the substructures responsible for the label prediction in the artificial experiment, even when the sample size is small. Furthermore, we use a real dataset to evaluate the effectiveness of the visualization. The basic idea of BayesGrad is not limited to graph-structured data and can be applied to other data types.
Twitter has been increasingly used for spreading messages about campaigns. Such campaigns try to gain followers through their Twitter accounts, influence the followers and spread messages through them. In this paper, we explore the relationship between followers’ sentiment towards the cam-paign topic and their rate of retweeting of messages gener-ated by the campaign. Our analysis with followers of mul-tiple social-media campaigns found statistical significant correlations between such sentiment and retweeting rate. Based on our analysis, we have conducted an online inter-vention study among the followers of different social-media campaigns. Our study shows that targeting followers based on their sentiment towards the campaign can give higher re-tweet rate than a number of other baseline approaches.
In machine learning, a nonparametric forecasting algorithm for time series data has been proposed, called the kernel spectral hidden Markov model (KSHMM). In this paper, we propose a technique for short-term wind-speed prediction based on KSHMM. We numerically compared the performance of our KSHMMbased forecasting technique to other techniques with machine learning, using wind-speed data offered by the National Renewable Energy Laboratory. Our results demonstrate that, compared to these methods, the proposed technique offers comparable or better performance. Keywords: Wind-Speed Prediction, Kernel Methods, Kernel Mean Embedding, Spectral Learning, Hidden Markov Models. 1. Introduction Wind energy is one of the most attractive renewable energy sources.