Recursive Structure.

Slides:



Advertisements
Similar presentations
Parsing with Compositional Vector Grammars Socher, Bauer, Manning, NG 2013.
Advertisements

Deep Learning in NLP Word representation and how to use it for Parsing
Deep Learning and Neural Nets Spring 2015
Duyu Tang, Furu Wei, Nan Yang, Ming Zhou, Ting Liu, Bing Qin
Distributed Representations of Sentences and Documents
Richard Socher Cliff Chiung-Yu Lin Andrew Y. Ng Christopher D. Manning
Deep Learning Neural Network with Memory (1)
TMpro: Transmembrane Helix Prediction using Amino Acid Properties and Latent Semantic Analysis Madhavi Ganapathiraju, N. Balakrishnan, Raj Reddy and Judith.
Kai Sheng-Tai, Richard Socher, Christopher D. Manning
Semantic Compositionality through Recursive Matrix-Vector Spaces
Predicting the dropouts rate of online course using LSTM method
Parsing Natural Scenes and Natural Language with Recursive Neural Networks INTERNATIONAL CONFERENCE ON MACHINE LEARNING (ICML 2011) RICHARD SOCHER CLIFF.
PROBABILISTIC GRAPH-BASED DEPENDENCY PARSING WITH CONVOLUTIONAL NEURAL NETWORK Zhisong Zhang, Hai Zhao and Lianhui QIN Shanghai Jiao Tong University
Learning to Answer Questions from Image Using Convolutional Neural Network Lin Ma, Zhengdong Lu, and Hang Li Huawei Noah’s Ark Lab, Hong Kong
Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation EMNLP’14 paper by Kyunghyun Cho, et al.
A Sentence Interaction Network for Modeling Dependence between Sentences Biao Liu, Minlie Huang Tsinghua University.
Medical Semantic Similarity with a Neural Language Model Dongfang Xu School of Information Using Skip-gram Model for word embedding.
Distributed Representations for Natural Language Processing
Deep Learning Methods For Automated Discourse CIS 700-7
Attention Model in NLP Jichuan ZENG.
Sentiment analysis using deep learning methods
End-To-End Memory Networks
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
CS 4501: Introduction to Computer Vision Computer Vision + Natural Language Connelly Barnes Some slides from Fei-Fei Li / Andrej Karpathy / Justin Johnson.
Deep Learning for Bacteria Event Identification
Sentence Modeling Representation of sentences is the heart of Natural Language Processing A sentence model is a representation and analysis of semantic.
Recursive Neural Networks
Speaker Classification through Deep Learning
Recurrent Neural Networks for Natural Language Processing
Van-Khanh Tran and Le-Minh Nguyen
Recurrent Neural Networks
Neural Machine Translation by Jointly Learning to Align and Translate
Progress Report WANG XUN 2015/10/02.
A Hierarchical Model of Reviews for Aspect-based Sentiment Analysis
Visualizing and Understanding Neural Models in NLP
References [1] - Y. LeCun, L. Bottou, Y. Bengio and P. Haffner, Gradient-Based Learning Applied to Document Recognition, Proceedings of the IEEE, 86(11): ,
Combining CNN with RNN for scene labeling (segmentation)
CSCI 5922 Neural Networks and Deep Learning: Image Captioning
Intelligent Information System Lab
Different Units Ramakrishna Vedantam.
Neural networks (3) Regularization Autoencoder
Deep learning and applications to Natural language processing
mengye ren, ryan kiros, richard s. zemel
Word2Vec CS246 Junghoo “John” Cho.
convolutional neural networkS
Distributed Representation of Words, Sentences and Paragraphs
Advanced Artificial Intelligence
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
A First Look at Music Composition using LSTM Recurrent Neural Networks
Recurrent Neural Networks
A Fast Unified Model for Parsing and Sentence Understanding
Tips for Training Deep Network
Understanding LSTM Networks
Word embeddings based mapping
Other Classification Models: Recurrent Neural Network (RNN)
Introduction to Natural Language Processing
Machine Translation(MT)
Vector Representation of Text
Presentation By: Eryk Helenowski PURE Mentor: Vincent Bindschaedler
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
Recurrent Neural Networks (RNNs)
Sentiment Classification
The experiments based on Recurrent Neural Networks
Baseline Model CSV Files Pandas DataFrame Sentence Lists
Recurrent Neural Networks
Deep learning: Recurrent Neural Networks CV192
Bidirectional LSTM-CRF Models for Sequence Tagging
Vector Representation of Text
CS249: Neural Language Model
Presentation transcript:

Recursive Structure

Application: Sentiment Analysis h0 f h1 f h2 f h3 f g sentiment x1 x2 x3 x4 word vector h3 g Word Sequence Recurrent Structure f Special case of recursive structure h1 same dimension h2 Recursive Structure How to stack function f is already determined f f x1 x2 x3 x4

Recursive Model word sequence: syntactic structure not very good not How to do it is out of the scope https://openreview.net/pdf?id=Skvgqgqxe Cons$tuency (phrase structure) word sequence: not very good

Recursive Model syntactic structure By composing the two meaning, what should the meaning be. not very good Dimension of word vector = |Z| Meaning of “very good” V(“very good”) Input: 2 X |Z|, output: |Z| Cons$tuency (phrase structure) f V(“not”) V(“very”) V(“good”) not very good

Recursive Model syntactic structure V(wA wB) ≠ V(wA) + V(wB) not very good “not”: neutral “good”: positive “not good”: negative Meaning of “very good” V(“very good”) Cons$tuency (phrase structure) f V(“not”) V(“very”) V(“good”) not very good

Recursive Model syntactic structure V(wA wB) ≠ V(wA) + V(wB) not very good “棒”: positive “好棒”: positive “好棒棒”: negative Meaning of “very good” V(“very good”) Cons$tuency (phrase structure) network f V(“not”) V(“very”) V(“good”) not very good

: “reverse” another input Recursive Model syntactic structure “not good” “not bad” not very good f f Meaning of “very good” “not” “good” “not” “bad” V(“very good”) : “reverse” another input Cons$tuency (phrase structure) f “not” V(“not”) V(“very”) V(“good”) not very good

: “emphasize” another input Recursive Model syntactic structure “very good” “very bad” not very good f f Meaning of “very good” “very” “good” “very” “bad” V(“very good”) : “emphasize” another input Cons$tuency (phrase structure) f “very” V(“not”) V(“very”) V(“good”) not very good

5 classes ( -- , - , 0 , + , ++ ) ref - output g Train both … V(“not very good”) f f g http://nlp.stanford.edu:8080/sentiment/rntnDemo.html Cons$tuency (phrase structure) f V(“not”) V(“very”) V(“good”) not very good

Recursive Neural Tensor Network 𝑝 W =𝜎 f Little interaction between a and b 𝑎 𝑏 𝑖,𝑗 𝑊 𝑖𝑗 𝑥 𝑖 𝑥 𝑗 W x xT Advantages Single layer can model complex interactions Disadvantages Tensor V is potentially gigantic Room for research on tensor networks? Does not appear to be well-explored Cubic parameter growth Highly expressive =𝜎 +

Experiments 5-class sentiment classification ( -- , - , 0 , + , ++ ) Demo: http://nlp.stanford.edu:8080/sentiment/rntnDemo.html

Experiments Dull 平淡 Socher, Richard, et al. "Recursive deep models for semantic compositionality over a sentiment treebank." Proceedings of the conference on empirical methods in natural language processing (EMNLP). Vol. 1631. 2013.

Matrix-Vector Recursive Network inherent meaning how it changes the others f = W 𝜎 = WM = = Richard Socher Brody Huval Christopher D. Manning Andrew Y. Ng EMNLP 2012 Informative? 𝑎 𝐴 𝑏 𝐵 Zero? −1 0 0 −1 ? Identity? not good

Tree LSTM Typical LSTM Tree LSTM ht-1 ht LSTM mt-1 mt xt h3 m3 LSTM h1

More Applications Sentence relatedness NN Recursive Neural Network Tai, Kai Sheng, Richard Socher, and Christopher D. Manning. "Improved semantic representations from tree-structured long short-term memory networks." arXiv preprint arXiv:1503.00075 (2015). Sentence 1 Sentence 2