This is a brief summary of paper for me to study and organize it, Modeling Rich Contexts for Sentiment Classification with LSTM (Huang et al., arXiv 2016) I read and studied.

They propose how to get representation with rich context in tweets which constitute a thread like conversation.

Huang et al., 2016

They used two types of LSTM, one is Word-level LSTM, the other is Tweet-level LSTM as above figure.

The word-level encode a tweet into the fixed-size vector.

and the Tweet-level use the vectors from word-level LSTM as input.

In other word,

From Figure 2 we can see that in the word-level LSTM, the input is individual words.
The hidden state of the last word is taken as the representation of the tweet.
Each tweet in a thread will go through the word-level LSTM, and the t-th tweet in the thread will generate a tweet representation of zt which will be an input of the tweet-level LSTM.

And the extend input of tweet level LSTM with context featues which represent as binary value.

The additional context feature is as follows:

  • SameAuthor
  • Converstaion
  • SameHashtag
  • SameEmoji

Reference