Presentation RaviShankar
Presentation RaviShankar
Presentation RaviShankar
Networks in NLP
D Ravi Shankar
Experiment 1
Experiment 2
Transfer of parameters layer wise
Experiment 3
1) Transfer Learning for various amounts of target
data.
2) INIT vs MULT.
When to transfer from source to target?
Conclusions
● Transfer Learning is successful when we are dealing with
semantically similar tasks.
● Transfer Learning also depends on what layers we are
transferring.
● It is helpful when the target dataset is small and also helps
in faster convergence.
● INIT method is performing slightly better compared to MULT.
● Are we losing general information if the model is
trained on source data for best accuracy? The answer
seems to be NO.
Future Work
● Conduct experiments and analyse which part of
the network can be transferred depending on
the type of application.
● Find the similarity between the datasets
quantitatively.
● NMT system first reads the source sentence using an
encoder to build a ”context” vector, a sequence of numbers
that represents the sentence meaning; a decoder, then,
processes the sentence vector to emit a translation.
● The context vector thus contains sufficient lexical and
semantic information to fully reconstruct a sentence in
another language.
● Transfer this knowledge (”context vector”) from NMT and
use it in variety of applications like sentiment analysis,
paraphrase detection, etc and see how much it helps in
improving the performance on the target dataset.
References
1) Lili Mou, Zhao Meng, Rui Yan, Ge Li, Yan Xu,Lu Zhang, Zhi Jin.How
Transferable are NeuralNetworks in NLP Applications?. In Proceedings
ofthe 2016 Conference on Empirical Methods in Nat-ural Language
Processing (EMNLP), pages 478–489, 2016.
2) Zhilin Yang, Ruslan Salakhutdinov, William W.Cohen.Transfer Learning
for Sequence Taggingwith Hierarchical Recurrent Networks. ICLR 2017.
3) Seunghyun Yoon, Hyeongu Yun, Yuna Kim, Gyu-tae Park, Kyomin
Jung.Efficient Transfer Learn-ing Schemes for Personalized Language
Modelingusing Recurrent Neural Network.CoRR 2017, vol-ume:
abs/1701.03578.
4) Sinno Jialin Pan and Qiang Yang.A sur-vey on transfer learning.
IEEE Transactions onKnowledge and Data Engineering 2010,
22(10),13451359.
5) Ilya Sutskever, Oriol Vinyals, and Quoc VVLe.Sequence to
Sequence Learning with NeuralNetworks.In Proceedings of EMNLP
2014,pages17241734.
6) Kyunghyun Cho, Bart van Merrienboer, aglarGlehre, Dzmitry
Bahdanau, Fethi Bougares, Hol-ger Schwenk, and Yoshua
Bengio.Learning phraserepresentations using RNN encoder-
decoder forstatistical machine translation.In Advances inNeural
Information Processing Systems 2014,pages 31043112.
7) Engineering at Quora.
https://engineering.quora.com/Semantic-Question-Matching-with-Deep
-Learning
8) Sebastian Ruder blog.
http://ruder.io/transfer-learning/
Thank you