Interesting Websites
-
Papers with Code.
-
NLP Progress.
-
Ruder.io.
Books:
-
NLP Notes by Jacob Eisenstein (2018 draft).
-
Information Retrieval by Manning, Raghavan and Schutze (2009).
Longer list of papers:
-
(2003) A Neural Probabilistic Language Model (NNML) by Bengio, Ducharme, Vincent, Jauvin, J Machine Learning Research 3, p 1137-1155.
-
(2013) Efficient Estimation of Word Representations in Vector Space by Mikolov, Chen, Corrado & Dean, ICLR 2013.
-
(2013) Distributed Representations of Words and Phrases and their Compositionality by Mikolov, Sutskever, Chen, Corrado & Dean, NIPS 2013.
-
(2014) GloVe: Global Vectors for Word Representation by Pennington, Socher & Manning, EMNLP 2014.
—
GloVe Homepage
-
(2014) Distributed Representations of Sentences and Documents by Le and Mikolov, ICML 2014.
-
(2015) Skip-Thought Vectors by Kiros, Zhu, Salakhutdinov, Zemel, Torralba, Urtasun & Fidler, NIPS 2015.
—
GitHub — Blog article
-
(2018) An Efficient Framework For Learning Sentence Representations by Logeswaran & Lee, ICLR 2018
—
GitHub (S2V) — OpenReview.
-
(2018) Supervised Learning of Universal Sentence Representations from Natural Language Inference Data by Conneau, Kiela, Schwenk, Barrault & Bordes, EMNLP 2017
—
GitHub (InferSent)
—
GitHub (SentEval)
—
Blog article 1
—
Blog article 2
—
Blog article 3
-
(2018) Learning General Purpose Distributed Sentence Representations via Large Scale Multi-task Learning by Subramanian, Trischler, Bengio & Pal, ICLR 2018
—
GitHub
-
(2018) Universal Sentence Encoder by Cer, Yang, Kong, Hua, Limtiaco, John, Constant, Guajardo-Cespedes, Yuan, Tar, Sung, STrope & Kurzweil, EMNLP 2018
—
TFHub
—
CoLab
-
(2017) Learned in Translation: Contextualized Word Vectors by McCann, Bradbury, Xiong & Socher, NIPS 2017
—
GitHub
—
Blog article 1
—
Blog article 2
-
(2018) Deep contextualized word representations by Peters, Neumann, Iyyer, Gardner, Clark, Lee & Zettlemoyer, ACL 2018
—
Elmo
—
GitHub
-
(2018) Universal Language Model Fine-tuning for Text Classification (ULMFit) by Howard & Ruder, ACL 2018
—
GitHub
-
Improving Language Understanding by Generative Pre-Training (GPT) by Radford, Narasimhan, Salimans & Sutskever
—
GitHub
—
Blog article 1
-
Attention Is All You Need
—
GitHub
—
Blog Article (The Illustrated Transformer by J Alammar)
-
Generating Wikipedia by Summarizing Long Sequences
—
GitHub
-
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
—
GitHub
—
Blog article: BERT Illustrated
—
Blog article: Introduction to the World of BERT
—
Blog article 3: BERT Explained
—
Blog article: BERT Explained - FAQ
—
Blog article: BERT EXplained
-
XLNet (Google)
-
RoBERTa (Google / CMU)
-
DistilBERT (HuggihgFace)
-
CTRL (Salesforce)
-
GPT-2 (OpenAI)
-
ALBERT (Google)
-
Megatron (NVIDIA)
-
(2019) Release Strategies and the Social Impacts of Language Models by Solaiman et al.
Blog articles:
(2016) Word2Vec Tutorial - The Skip-Gram Model by Chris McCormick.
(2018) NLP's ImageNet moment has arrived
(2019) Visualizing A Neural Machine Translation Model by Jay Alammar. Referencd by this lecture by Lex Fridman.
Deep Learning State of the Art (2020) by Lex Fridman.
Over 200 of the Best Machine Learning, NLP, and Python Tutorials — 2018 Edition by Robbie Allen
Software:
Write with Transformer by HuggingFace.
Dialog Systems:
(2019) Transferable Multi-Domain State Generator for Task-Oriented Dialogue Systems by Wu, Madotto, Hosseini-Asl, Xiong, Socher & Fung, ACL 2019).
(2019) by Rajani, McCann & Socher, ACL 2019.