Goto

Collaborating Authors

 variational model



Visual Adversarial Imitation Learning using Variational Models

Neural Information Processing Systems

Reward function specification, which requires considerable human effort and iteration, remains a major impediment for learning behaviors through deep reinforcement learning. In contrast, providing visual demonstrations of desired behaviors presents an easier and more natural way to teach agents. We consider a setting where an agent is provided a fixed dataset of visual demonstrations illustrating how to perform a task, and must learn to solve the task using the provided demonstrations and unsupervised environment interactions. This setting presents a number of challenges including representation learning for visual observations, sample complexity due to high dimensional spaces, and learning instability due to the lack of a fixed reward or learning signal. Towards addressing these challenges, we develop a variational model-based adversarial imitation learning (V-MAIL) algorithm. The model-based approach provides a strong signal for representation learning, enables sample efficiency, and improves the stability of adversarial training by enabling on-policy learning. Through experiments involving several vision-based locomotion and manipulation tasks, we find that V-MAIL learns successful visuomotor policies in a sample-efficient manner, has better stability compared to prior work, and also achieves higher asymptotic performance. We further find that by transferring the learned models, V-MAIL can learn new tasks from visual demonstrations without any additional environment interactions. All results including videos can be found online at https://sites.google.com/view/variational-mail


Visual Adversarial Imitation Learning using Variational Models

Neural Information Processing Systems

Reward function specification, which requires considerable human effort and iteration, remains a major impediment for learning behaviors through deep reinforcement learning.


EQNN: Enhanced Quantum Neural Network

Chen, Abel C. H.

arXiv.org Artificial Intelligence

With the maturation of quantum computing technology, research has gradually shifted towards exploring its applications. Alongside the rise of artificial intelligence, various machine learning methods have been developed into quantum circuits and algorithms. Among them, Quantum Neural Networks (QNNs) can map inputs to quantum circuits through Feature Maps (FMs) and adjust parameter values via variational models, making them applicable in regression and classification tasks. However, designing a FM that is suitable for a given application problem is a significant challenge. In light of this, this study proposes an Enhanced Quantum Neural Network (EQNN), which includes an Enhanced Feature Map (EFM) designed in this research. This EFM effectively maps input variables to a value range more suitable for quantum computing, serving as the input to the variational model to improve accuracy. In the experimental environment, this study uses mobile data usage prediction as a case study, recommending appropriate rate plans based on users' mobile data usage. The proposed EQNN is compared with current mainstream QNNs, and experimental results show that the EQNN achieves higher accuracy with fewer quantum logic gates and converges to the optimal solution faster under different optimization algorithms.


Visual Adversarial Imitation Learning using Variational Models

Neural Information Processing Systems

Reward function specification, which requires considerable human effort and iteration, remains a major impediment for learning behaviors through deep reinforcement learning. In contrast, providing visual demonstrations of desired behaviors presents an easier and more natural way to teach agents. We consider a setting where an agent is provided a fixed dataset of visual demonstrations illustrating how to perform a task, and must learn to solve the task using the provided demonstrations and unsupervised environment interactions. This setting presents a number of challenges including representation learning for visual observations, sample complexity due to high dimensional spaces, and learning instability due to the lack of a fixed reward or learning signal. Towards addressing these challenges, we develop a variational model-based adversarial imitation learning (V-MAIL) algorithm.


A Differentiable Partially Observable Generalized Linear Model with Forward-Backward Message Passing

Li, Chengrui, Li, Weihan, Wang, Yule, Wu, Anqi

arXiv.org Artificial Intelligence

The partially observable generalized linear model (POGLM) is a powerful tool for understanding neural connectivity under the assumption of existing hidden neurons. With spike trains only recorded from visible neurons, existing works use variational inference to learn POGLM meanwhile presenting the difficulty of learning this latent variable model. There are two main issues: (1) the sampled Poisson hidden spike count hinders the use of the pathwise gradient estimator in VI; and (2) the existing design of the variational model is neither expressive nor time-efficient, which further affects the performance. For (1), we propose a new differentiable POGLM, which enables the pathwise gradient estimator, better than the score function gradient estimator used in existing works. For (2), we propose the forward-backward message-passing sampling scheme for the variational model. Comprehensive experiments show that our differentiable POGLMs with our forward-backward message passing produce a better performance on one synthetic and two real-world datasets. Furthermore, our new method yields more interpretable parameters, underscoring its significance in neuroscience.


Learning Variational Models with Unrolling and Bilevel Optimization

Brauer, Christoph, Breustedt, Niklas, de Wolff, Timo, Lorenz, Dirk A.

arXiv.org Machine Learning

In this paper we consider the problem of learning variational models in the context of supervised learning via risk minimization. Our goal is to provide a deeper understanding of the two approaches of learning of variational models via bilevel optimization and via algorithm unrolling. The former considers the variational model as a lower level optimization problem below the risk minimization problem, while the latter replaces the lower level optimization problem by an algorithm that solves said problem approximately. Both approaches are used in practice, but unrolling is much simpler from a computational point of view. To analyze and compare the two approaches, we consider a simple toy model, and compute all risks and the respective estimators explicitly. We show that unrolling can be better than the bilevel optimization approach, but also that the performance of unrolling can depend significantly on further parameters, sometimes in unexpected ways: While the stepsize of the unrolled algorithm matters a lot (and learning the stepsize gives a significant improvement), the number of unrolled iterations plays a minor role.


Weakly supervised segmentation with point annotations for histopathology images via contrast-based variational model

Zhang, Hongrun, Burrows, Liam, Meng, Yanda, Sculthorpe, Declan, Mukherjee, Abhik, Coupland, Sarah E, Chen, Ke, Zheng, Yalin

arXiv.org Artificial Intelligence

Image segmentation is a fundamental task in the field of imaging and vision. Supervised deep learning for segmentation has achieved unparalleled success when sufficient training data with annotated labels are available. However, annotation is known to be expensive to obtain, especially for histopathology images where the target regions are usually with high morphology variations and irregular shapes. Thus, weakly supervised learning with sparse annotations of points is promising to reduce the annotation workload. In this work, we propose a contrast-based variational model to generate segmentation results, which serve as reliable complementary supervision to train a deep segmentation model for histopathology images. The proposed method considers the common characteristics of target regions in histopathology images and can be trained in an end-to-end manner. It can generate more regionally consistent and smoother boundary segmentation, and is more robust to unlabeled `novel' regions. Experiments on two different histology datasets demonstrate its effectiveness and efficiency in comparison to previous models.


Design Amortization for Bayesian Optimal Experimental Design

Kennamer, Noble, Walton, Steven, Ihler, Alexander

arXiv.org Artificial Intelligence

Bayesian optimal experimental design is a sub-field of statistics focused on developing methods to make efficient use of experimental resources. Any potential design is evaluated in terms of a utility function, such as the (theoretically well-justified) expected information gain (EIG); unfortunately however, under most circumstances the EIG is intractable to evaluate. In this work we build off of successful variational approaches, which optimize a parameterized variational model with respect to bounds on the EIG. Past work focused on learning a new variational model from scratch for each new design considered. Here we present a novel neural architecture that allows experimenters to optimize a single variational model that can estimate the EIG for potentially infinitely many designs. To further improve computational efficiency, we also propose to train the variational model on a significantly cheaper-to-evaluate lower bound, and show empirically that the resulting model provides an excellent guide for more accurate, but expensive to evaluate bounds on the EIG. We demonstrate the effectiveness of our technique on generalized linear models, a class of statistical models that is widely used in the analysis of controlled experiments. Experiments show that our method is able to greatly improve accuracy over existing approximation strategies, and achieve these results with far better sample efficiency.


Explainable bilevel optimization: an application to the Helsinki deblur challenge

Bonettini, Silvia, Franchini, Giorgia, Pezzi, Danilo, Prato, Marco

arXiv.org Artificial Intelligence

In general, H is a structured matrix defined in such a way that the product Hu corresponds to a convolution between the image u and a given kernel h representing the Point Spread Function (PSF) of the imaging system employed to measure the data. The deblurring (or deconvolution) problem consists in finding an approximation of g, given the blurred image f and, possibly, some information on the system PSF. If the blurring kernel h, underlying the matrix H, is completely unknown and it has to be inferred together with g, the resulting problem is a blind deconvolution one [32]. Since the PSF h usually represents a low-pass filter, the matrix H is, at best, very ill conditioned and directly solving the inverse problem Hu = f, even when it is feasible, leads to unmeaningful solutions. On the other side, the variational approach consists in designing and solving an optimization problem whose solutions are a good approximation of the unknown image g. In general, a variational model is the set composed by the objective function, i.e., the function to be minimized, and the possible constraints. In the variational models arising in image restoration applications, the objective function, called also energy functional, encompasses different kinds of information: the nature of the noise introduced in the acquisition process, geometrical and/or analytical properties on the image content and physical constraints on the pixel values. Usually, in all image reconstruction problems, and more generally inverse problems, the energy functional, besides the data, depends on a set of parameters; they may simply reduce to tuning parameters balancing the relative weights of the different terms in the functional but can also represent more complicate structures of the functionals themselves.