Skip to content

Files

Latest commit

 

History

History

2_intermediate

Credits

This is a remix from https://github.com/DeepLearningDTU/02456-deep-learning

Content

This is the next steps in learning how to use PyTorch. The purpose is to get the audience to get the audience familiar with building machine learning models in PyTorch.

NOTE In these labs cuda is used extensively. If you have a cuda enabled machine, read the README.md in the root of this repo on how to use nvidia-docker.

Lab 1: Feed Forward Network

In this lab you will implement a simple feed forward neural network in PyTorch.

Reading material

1.1-FFN

Train a FFN classifier to solve the half moon problem.

1.2-FFN

Train a FFN classifier to recognize handwritten digits.

Lab 2: Convolutional Neural Network

In this lab you will implement a convolutional neural network in PyTorch.

Reading material

2.1-CNN

Train a CNN classifier to recognize handwritten digits.

2.2-CNN

Use the cluttered MNIST dataset along with a Spatial Transformer Network to make a CNN classifier to recognize handwritten digits.

This lab contains first real use of cuda support.

Lab 3: Autoencoders

Autoencoders and MNIST.

Reading material

3.1-AE

Train an autoencoder on MNIST to encode a target given some input.

Lab 4: Recurrent Neural Networks and Sequences

This lab shows how sequences can be used with recurrent neural networks to learn to classify text.

Reading material

4.1-Sequences

Train a bag of words model to classify positive and negative sentences.

4.2-RNN

Train a recurrent neural network to classify positive and negative sentences. Note this lab is very slow without a GPU. See README.md in root folder.

Lab 5: Leaf Classification with Kaggle

Combine all previous labs (FFN, CNN, RNN) to compete in the leaf classification challenge.

Reading material

Nothing.

5.1-Kaggle

Train different models and submit the results to Kaggle.

Lab 6: Final exam

Test your newly acquired skills on the most notorious classification challenges: Text Entailment.

Reading material

6.1-final_exam_SNLI

Build data loader, model, loss function, optimizer and training loop from scratch. Notice that replicating the BCN model is a difficult task and may take a significant amount of time.