Starting from:

$29.99

Assignment 4 – RNN, LSTM for language modeling and classification

CS 505– Assignment 4 (100 pts, BONUS: 10 pts) – RNN, LSTM for language modeling and classification

In this assignment, you will learn about text classification and language modeling using RNN and LSTM, and use python
framework for deep learning such as pytorch, which are popular in NLP. You have 2 weeks to finish this particular assignment.
Submit in Blackboard by 11:59PM EST, March 27.
–Please indicate names of those you collaborate with.
–Every late day will reduce your score by 20
–After 2 days (i.e., if you submit on the 3rd day after due date), it will be marked 0.
Submit your (1) code/Jupyter notebook and (2) write up in one zip file.
When necessary, you must show how you derive your answer
Problem 1. Neural Models (32 pts)
1. (2 pts) In a news framing classification task, where you have 5 frames and your model predicts each of the frames with
equal probability for an article, what is the cross entropy loss of the article in this case?
2. (2 pts) Suppose during training of your neural model you realize that your training loss remains high. Mention some of
the ways you can reduce this underfitting of your neural network.
3. (2 pts) After you do many changes to your neural network, you now realize that your training loss is much lower than
your validation loss. Mention some of the ways you can reduce this overfitting of your neural network.
4. (2 pts) What is good about setting a large batch size for training? How about a small batch size?
5. (3 pts) How can an RNN be used for detecting toxic spans (spans of words containing toxic language) in a social media
comment? Specifically, what should be the input to the RNN at each time step t? How many outputs (i.e., yˆ) are produced
given a comment containing n words? What is each yˆ
(t) a probability distribution over?
6. (3 pts) How about using RNNs for language modeling? Given a start word token as input at time step 1, what should
be the input to the RNN at each time step t > 1? How many outputs are produced? What is each yˆ
(t) a probability
distribution over?
7. (3 pts) How about using RNNs for frame classification? Given an article containing n words as input, what should be
the input to the RNN at each time step t? How many outputs are produced? What is each yˆ
(t) a probability distribution
over?
8. (2 pts) What is the main advantage of using RNNs for frame classification over feed forward neural network?
9. (3 pts) What is the disadvantage of RNN when used to classify the sentiment of a very long tweet like this? “I am not
sure I want this phone. It’s too big to fit in my back pocket. I put it in and accidentally sat on it and now it’s bent. I’m
very disappointed. I’m now the proud owner of bendy iPhone6. Very proud.” What is the appropriate sentiment for this
tweet? And what would the RNN classify it as?
10. How about LSTM? Given this formulation of LSTM: ft = σ(Wfxt + Ufht−1 + bf ) (forget gate), it = σ(Wixt +
Uiht−1 + bi) and Cˆ
t = tanh(WC xt + UC ht−1 + bC ) (input gate), Ct = ft ∗ Ct−1 + it ∗ Cˆ
t (update gate), and
ot = σ(Woxt + Uoht−1 + bo) and ht = ot ∗ tanh(Ct):
(a) (4 pts) derive the formulation of ∂J
UC
for two time steps t and t − 1 in terms of ∂J
∂ht
,
∂ht
∂Ct
,
∂Ct
∂UC
,
∂Ct
∂Ct−1
,
∂Ct−1
∂UC
,
∂ht
∂ht−1
,
and ∂ht−1
∂UC
.
(b) (2 pts) which part of ∂J
UC
reduces the effect of the vanishing gradient problem in RNNs?
(c) (2 pts) How does this help classify the correct sentiment of the tweet above?
(d) (2 pts) Instead of using the last hidden state of LSTM to classify the tweet, what other ways we can do to improve
the performance of this sentiment classification?
1
Problem 2. LSTM for language modeling (36 points)
1. (5 pts) Follow the tutorial in here to train a word-level LSTM language modeling. Train the language model on texts
from the file prideAndPrejudice.txt. Before using it to train the language model, you need to first sentence segment, then
tokenize, then lower case each line of the file using Spacy. Append start-of-sentence token ’<s>’ and end-of-sentence
’</s>’ token to each sentence and put each sentence in its own line. Use only words that appear more than once in this
corpus and assign UNK tokens for the rest; you may also need to pad sentences that are shorter than 5 (see here in cell
10-12 for adding unknown: UNK token and padding: PAD token to your vocabulary). Train the language model and
save the trained model (see here). Generate 10 examples of text from it, starting from ’<s>’ token and ending at ’</s>’
token.
2. (5 pts) Compute and report the perplexity of the saved model on test 1.txt file. Note that the test files are already preprocessed.
3. (5 pts) Train the language model as before, but with input sequence lengths of 25 (currently, it’s inputs are of length 5).
You may need to pad some of the shorter sentences to length 25. Save your trained model. Generate 10 examples of text
from it, starting from ’<s>’ token and ending at ’</s>’ token. Are there differences from the generated examples from
2.1?
4. (1 pts) Compute and report the perplexity of this saved model on test 1.txt file.
5. (1 pts) Use the better language model (the one with the lower perplexity on test 1.txt) to compute and report the perplexity
on test 2.txt. Note that the test files are already pre-processed.
6. (5 pts) Train the better language model as before but start with pre-trained Glove6B 100d embeddings (see here on how
to incorporate pretrained embeddings in your LSTM model). This time, use all your words, even those occurring only
once in the corpus. Only assign UNK token to words that are not in Glove vocabulary and initialize random vectors in
the embedding matrix for the UNK, <s>, </s>, and PAD tokens. Save your trained model. Generate 10 examples of text
from it, starting from ’<s>’ token and ending at ’</s>’ token. Are there differences from the generated examples from
before?
7. (1 pts) Compute and report the perplexity of this saved model on test 1.txt file.
8. (2 pts) Train a language model with input sequence lengths of 5 as before (Question 2.1) on texts from tweet.txt. Note that
this file is already pre-processed. Save your trained model. Generate 10 examples of text from it, starting from ’<s>’
token and ending at ’</s>’ token.
9. (1 pts) Compute and report the perplexity of this saved model on test 2.txt file.
10. (2 pts) Train a language model with input sequence lengths of 15 on texts from tweet.txt. Save your trained model.
Generate 10 examples of text from it, starting from ’<s>’ token and ending at ’</s>’ token. Are there differences from
the generated examples from 2.8?
11. (1 pts) Compute and report the perplexity of this saved model on test 2.txt file.
12. (1 pts) Use the better language model (the one with the lower perplexity on test 2.txt) to compute and report the perplexity
on test 1.txt.
13. (2 pts) Train the better language model on tweet.txt but starting from pre-trained Glove6B 100d embeddings like in 2.6.
Save your trained model. Generate 10 examples of text from it, starting from ’<s>’ token and ending at ’</s>’ token.
14. (1 pts) Compute and report the perplexity of this saved model on test 2.txt file.
15. (2 pts) Train the better language model on tweet.txt but starting from pre-trained GloveTwitter 100d like in 2.6. Save
your trained model. Generate 10 examples of text from it, starting from ’<s>’ token and ending at ’</s>’ token.
16. (1 pts) Compute and report the perplexity of this saved model on test 2.txt file.
Problem 3. LSTM for classification (32 points, BONUS: 10 pts)
1. (5 pts) Follow the tutorial here on how to build LSTM model for sentiment classification. Modify the tutorial to train on
your tweet sentiment data (sentiment-train.csv) and test on test data (sentiment-test.csv) from HW3 (modify the tutorial
so that the train data is not split into train and validation). Compute and report the accuracy on the test data.
2
2. (2 pts) Modify the model from 3.1 to use GRU. Compute and report the accuracy on the test data.
3. (5 pts) Modify the model from 3.1 to use bidirectional LSTM. Compute and report the accuracy on the test data.
4. (2 pts) Modify the model from 3.1 to use bidirectional GRU. Compute and report the accuracy on the test data.
5. (5 pts) Pick the best model so far and train the model starting from pretrained GloveTwitter 100d. Compute and report
the accuracy on the test data.
6. (10 pts) Using your best model so far, conduct a 5-fold (stratified) cross validation on your training data and a grid
search to pick the best hidden size (try 128 or 512) and embedding size (try 100 or 400). Compute and report the average
accuracies for each of the choice combination.
7. (3 pts) Train the model on all your training data using the best combination of hyperparameters you find in 2.6. Compute
and report the accuracy on the test data.
8. (BONUS: 10 pts) Train your best model using the hyperparameter from 2.6 on all the sentiment140 data. Compute and
report the accuracy on the test data from HW3 (i.e., sentiment-test.csv) .
3

More products