This is a course on natural language processing.
-
Lecturer: Felipe Bravo-Marquez
-
TAs: Matías Rojas, Mauricio Araneda y Jesús Pérez-Martín
-
Lectures: Tuesday 14:30 - 16:00, Thursday 14:30 - 16:00
-
Course Program (in Spanish)
The neural network-related topics of the course are taken from the book of Yoav Goldberg: Neural Network Methods for Natural Language Processing. The non-neural network topics (e.g., grammars, HMMS) are taken from the course of Michael Collins.
- Introduction to Natural Language Processing | (tex source file), video 1, video 2
- Vector Space Model and Information Retrieval | (tex source file), video 1, video 2
- Language Models (slides by Michael Collins), notes, video 1, video 2, video 3, video 4
- Text Classification and Naive Bayes (slides by Dan Jurafsky), notes, video 1, video 2, video 3
- Linear Models | (tex source file), video 1, video 2, video 3, video 4
- Neural Networks | (tex source file), video 1, video 2, video 3, video 4
- Word Vectors | (tex source file) video 1, video 2, video 3
- Tagging, and Hidden Markov Models (slides by Michael Collins), notes, video 1, video 2, video 3, video 4
- MEMMs and CRFs | (tex source file), notes 1, notes 2, video 1, video 2, video 3
- Convolutional Neural Networks | (tex source file), video
- Recurrent Neural Networks | (tex source file), video 1, video 2, video 3
- Sequence to Sequence Models, Attention, and the Transformer | (tex source file), video 1, video 2, video 3
- Contextual Words Representations (slides by Chris Manning) video 1, video 2, video 3, notes
- Constituency Parsing slides 1, slides 2, slides 3, slides 4 (slides by Michael Collins), notes 1, notes 2, videos 1, videos 2, videos 3, videos 4
- Recursive Networks and Paragraph Vectors | (tex source file)
- NLTK: Natural Language Toolkit
- Gensim
- spaCy: Industrial-strength NLP
- Torchtext
- AllenNLP: Open source project for designing deep leaning-based NLP models
- Transformers: a library of state-of-the-art pre-trained models for Natural Language Processing (NLP)
- Stanza - A Python NLP Library for Many Human Languages
- FlairNLP: A very simple framework for state-of-the-art Natural Language Processing (NLP)
- WEFE: The Word Embeddings Fairness Evaluation Framework
- WhatLies: A library that tries help you to understand. "What lies in word embeddings?"
- LASER:a library to calculate and use multilingual sentence embeddings
- Speech and Language Processing (3rd ed. draft) by Dan Jurafsky and James H. Martin.
- Michael Collins' NLP notes.
- A Primer on Neural Network Models for Natural Language Processing by Joav Goldberg.
- Natural Language Understanding with Distributed Representation by Kyunghyun Cho
- Natural Language Processing Book by Jacob Eisenstein
- NLTK book
- Embeddings in Natural Language Processing by Mohammad Taher Pilehvar and Jose Camacho-Collados
- Dive into Deep Learning Book
- CS224n: Natural Language Processing with Deep Learning, Stanford course
- Deep Learning in NLP: slides by Horacio Rodríguez
- David Bamman NLP Slides @Berkley
- CS 521: Statistical Natural Language Processing by Natalie Parde, University of Illinois
- 10 Free Top Notch Natural Language Processing Courses
- Natural Language Processing MOOC videos by Dan Jurafsky and Chris Manning, 2012
- Natural Language Processing MOOC videos by Michael Collins, 2013
- Natural Language Processing with Deep Learning by Chris Manning and Richard Socher, 2017
- CS224N: Natural Language Processing with Deep Learning | Winter 2019
- Computational Linguistics I by Jordan Boyd-Graber University of Maryland
- Visualizing and Understanding Recurrent Networks
- BERT Research Series by Chris McCormick
- Successes and Challenges in Neural Models for Speech and Language - Michael Collins
- More on Transforemers: BERT and Friends by Jorge Pérez
- ACL Portal
- NLP-progress: Repository to track the progress in Natural Language Processing (NLP)
- NLP News By Sebastian Ruder
- Corpora Mailing List
- Real World NLP Book: AllenNLP tutorials
- Attention is all you need explained
- The Illustrated Transformer: a very illustrative blog post about the Transformer
- ELMO explained
- BERT exaplained
- Better Language Models and Their Implications OpenAI Blog
- RNN effectiveness
- SuperGLUE: an benchmark of Natural Language Understanding Tasks
- decaNLP The Natural Language Decathlon: a benchmark for studying general NLP models that can perform a variety of complex, natural language tasks.
- Chatbot and Related Research Paper Notes with Images
- XLNet Explained
- Ben Trevett's torchtext tutorials
- PLMpapers: a collection of papers about Pre-Trained Language Models
- The Illustrated GPT-2 (Visualizing Transformer Language Models)
- Linguistics, NLP, and Interdisciplinarity Or: Look at Your Data, by Emily M. Bender
- The State of NLP Literature: Part I, by Saif Mohammad
- From Word to Sense Embeddings:A Survey on Vector Representations of Meaning
- 10 ML & NLP Research Highlights of 2019 by Sebastian Ruder
- Towards a Conversational Agent that Can Chat About…Anything
- The Super Duper NLP Repo: a collection of Colab notebooks covering a wide array of NLP task implementations
- The Big Bad NLP Database, a collection of nearly 300 well-organized, sortable, and searchable natural language processing datasets
- A Primer in BERTology: What we know about how BERT works
- How Self-Attention with Relative Position Representations works
- Deep Learning Based Text Classification: A Comprehensive Review
- Teaching NLP is quite depressing, and I don't know how to do it well by Yoav Goldberg