cnn-f
Appendix
In this section, we provide proofs for Proposition 2.1.B. Inthe proof, we inherit the notations that weuseforprovingTheorem2.1. The instance normalization that we incorporate into the DGM is not the same as the instance normalization that is typically used in image stylization [35]. CNN-F-5 significantly improves the robustness of CNN. CNN-F achieves higher accuracy on MNIST than CNN for under both standard training and adversarial training.
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > Jordan (0.04)
- Information Technology (0.68)
- Health & Medicine > Therapeutic Area > Neurology (0.47)
Appendix A Inference in the Generative Model
A.1 Generative model We choose the deconvolutional generative model (DGM) [25] as the generative feedback in CNN-F. The graphical model of the DGM is shown in Figure 2 (middle). In this section, we provide proofs for Theorem 2.1. Without loss of generality, we consider a DGM that has the following architecture. Lemma A.1 shows that logits output from the corresponding CNN of the DGM is proportional to the inner product of generated image and input image plus Lemma A.1 to show that CNN performs Bayesian inference in the DGM.
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > Jordan (0.04)
- Information Technology (0.68)
- Health & Medicine > Therapeutic Area > Neurology (0.47)
- Law > Litigation (0.42)
- Health & Medicine > Therapeutic Area > Neurology (0.41)
Neural Networks with Recurrent Generative Feedback
Huang, Yujia, Gornet, James, Dai, Sihui, Yu, Zhiding, Nguyen, Tan, Tsao, Doris Y., Anandkumar, Anima
Neural networks are vulnerable to input perturbations such as additive noise and adversarial attacks. In contrast, human perception is much more robust to such perturbations. The Bayesian brain hypothesis states that human brains use an internal generative model to update the posterior beliefs of the sensory input. This mechanism can be interpreted as a form of self-consistency between the maximum a posteriori (MAP) estimation of an internal generative model and the external environment. Inspired by such hypothesis, we enforce self-consistency in neural networks by incorporating generative recurrent feedback. We instantiate this design on convolutional neural networks (CNNs). The proposed framework, termed Convolutional Neural Networks with Feedback (CNN-F), introduces a generative feedback with latent variables to existing CNN architectures, where consistent predictions are made through alternating MAP inference under a Bayesian framework. In the experiments, CNN-F shows considerably improved adversarial robustness over conventional feedforward CNNs on standard benchmarks.
- North America > United States > California (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Asia > Middle East > Jordan (0.04)
- Government > Military (0.48)
- Health & Medicine > Therapeutic Area > Neurology (0.46)
- Information Technology > Security & Privacy (0.34)
Feature-Extracting Functions for Neural Logic Rule Learning
Gupta, Shashank, Robles-Kelly, Antonio
In this paper, we present a method aimed at integrating domain knowledge abstracted as logic rules into the predictive behaviour of a neural network using feature extracting functions for visual sentiment analysis. We combine the declarative first-order logic rules which represent the human knowledge in a logically-structured format making use of feature-extracting functions. These functions are embodied as programming functions which can represent, in a straightforward manner, the applicable domain knowledge as a set of logical instructions and provide a cumulative set of probability distributions of the input data. These distributions can then be used during the training process in a mini-batch strategy. In contrast with other neural logic approaches, the programmatic nature in practice of these functions do not require any kind of special mathematical encoding, which makes our method very general in nature. We also illustrate the utility of our method for sentiment analysis and compare our results to those obtained using a number of alternatives elsewhere in the literature.
- Information Technology > Communications > Social Media (1.00)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (1.00)