On Education Natural Language Processing with Deep Learning in Python - all courses

#artificialintelligence

Understand and implement word2vec Understand the CBOW method in word2vec Understand the skip-gram method in word2vec Understand the negative sampling optimization in word2vec Understand and implement GloVe using gradient descent and alternating least squares Use recurrent neural networks for parts-of-speech tagging Use recurrent neural networks for named entity recognition Understand and implement recursive neural networks for sentiment analysis Understand and implement recursive neural tensor networks for sentiment analysis Install Numpy, Matplotlib, Sci-Kit Learn, Theano, and TensorFlow (should be extremely easy by now) Understand backpropagation and gradient descent, be able to derive and code the equations on your own Code a recurrent neural network from basic primitives in Theano (or Tensorflow), especially the scan function Code a feedforward neural network in Theano (or Tensorflow) Helpful to have experience with tree algorithms In this course we are going to look at advanced NLP. Previously, you learned about some of the basics, like how many NLP problems are just regular machine learning and data science problems in disguise, and simple, practical methods like bag-of-words and term-document matrices. These allowed us to do some pretty cool things, like detect spam emails, write poetry, spin articles, and group together similar words. In this course I'm going to show you how to do even more awesome things. We'll learn not just 1, but 4 new architectures in this course.


Modeling Compositionality with Multiplicative Recurrent Neural Networks

arXiv.org Machine Learning

We present the multiplicative recurrent neural network as a general model for compositional meaning in language, and evaluate it on the task of fine-grained sentiment analysis. We establish a connection to the previously investigated matrix-space models for compositionality, and show they are special cases of the multiplicative recurrent net. Our experiments show that these models perform comparably or better than Elman-type additive recurrent neural networks and outperform matrix-space models on a standard fine-grained sentiment analysis corpus. Furthermore, they yield comparable results to structural deep models on the recently published Stanford Sentiment Treebank without the need for generating parse trees.


Deep Recursive Neural Networks for Compositionality in Language

Neural Information Processing Systems

Recursive neural networks comprise a class of architecture that can operate on structured input. They have been previously successfully applied to model compositionality in natural language using parse-tree-based structural representations. Even though these architectures are deep in structure, they lack the capacity for hierarchical representation that exists in conventional deep feed-forward networks as well as in recently investigated deep recurrent neural networks. In this work we introduce a new architecture --- a deep recursive neural network (deep RNN) --- constructed by stacking multiple recursive layers. We evaluate the proposed model on the task of fine-grained sentiment classification. Our results show that deep RNNs outperform associated shallow counterparts that employ the same number of parameters. Furthermore, our approach outperforms previous baselines on the sentiment analysis task, including a multiplicative RNN variant as well as the recently introduced paragraph vectors, achieving new state-of-the-art results. We provide exploratory analyses of the effect of multiple layers and show that they capture different aspects of compositionality in language.


Primer on Neural Network Models for Natural Language Processing - Machine Learning Mastery

#artificialintelligence

Deep learning is having a large impact on the field of natural language processing. But, as a beginner, where do you start? Both deep learning and natural language processing are huge fields. What are the salient aspects of each field to focus on and which areas of NLP is deep learning having the most impact? In this post, you will discover a primer on deep learning for natural language processing.


10 Deep Learning Terms Explained in Simple English - AYLIEN

#artificialintelligence

Deep Learning is a new area of Machine Learning research that has been gaining significant media interest owing to the role it is playing in artificial intelligence applications like image recognition, self-driving cars and most recently the AlphaGo vs. Lee Sedol matches. Recently, Deep Learning techniques have become popular in solving traditional Natural Language Processing problems like Sentiment Analysis. For those of you that are new to the topic of Deep Learning, we have put together a list of ten common terms and concepts explained in simple English, which will hopefully make them a bit easier to understand. We've done the same in the past for Machine Learning and NLP terms, which you might also find interesting. In the human brain, a neuron is a cell that processes and transmits information.