Goto

Collaborating Authors

 Raghavan, Aditya


Rapid optimization in high dimensional space by deep kernel learning augmented genetic algorithms

arXiv.org Artificial Intelligence

Pacific Northwest National Laboratory, Richland, WA Exploration of complex high-dimensional spaces presents significant challenges in fields such as molecular discovery, process optimization, and supply chain management. Genetic Algorithms (GAs), while offering significant power for creating new candidates' spaces, often entail high computational demands due to the need for evaluation of each new proposed solution. On the other hand, Deep Kernel Learning (DKL) efficiently navigates the spaces of preselected candidate structures but lacks generative capabilities. This study introduces an approach that amalgamates the generative power of GAs to create new candidates with the efficiency of DKL-based surrogate models to rapidly ascertain the behavior of new candidate spaces. This DKL-GA framework can be further used to build Bayesian Optimization (BO) workflows. We demonstrate the effectiveness of this approach through the optimization of the FerroSIM model, showcasing its broad applicability to diverse challenges, including molecular discovery and battery charging optimization. Manufacturing and chemical engineering often require complex resource allocation and control problems. Optimization in molecular spaces is crucial because it enables the discovery of new compounds with desired properties, leading to breakthroughs in various fields such as pharmaceuticals, materials science, and energy storage. Efficiently exploring molecular spaces allows researchers to identify novel molecules that can serve as effective drugs, advanced materials with superior properties, or catalysts for chemical reactions. By optimizing molecular spaces, we can uncover hidden relationships and patterns that lead to more efficient and targeted experimentation, reducing costs and time associated with traditional trial-and-error methods.


Hidden Markov Models with Random Restarts vs Boosting for Malware Detection

arXiv.org Artificial Intelligence

Effective and efficient malware detection is at the forefront of research into building secure digital systems. As with many other fields, malware detection research has seen a dramatic increase in the application of machine learning algorithms. One machine learning technique that has been used widely in the field of pattern matching in general-and malware detection in particular-is hidden Markov models (HMMs). HMM training is based on a hill climb, and hence we can often improve a model by training multiple times with different initial values. In this research, we compare boosted HMMs (using AdaBoost) to HMMs trained with multiple random restarts, in the context of malware detection. These techniques are applied to a variety of challenging malware datasets. We find that random restarts perform surprisingly well in comparison to boosting. Only in the most difficult "cold start" cases (where training data is severely limited) does boosting appear to offer sufficient improvement to justify its higher computational cost in the scoring phase.