Attention-based Multi-hop Recurrent Neural Network (AMRNN) Model

MM 03/12/2018


Fig. 2 The overall structure of the proposed Attention-based Multi-hop Recurrent Neural Network (AMRNN) model.

Fig.2 shows the overall structure of the AMRNN model.

The input of model includes the transcriptions of an audio story, a question and four answer choices, all represented as word sequences. The word sequence of the input question is represented as a question vector V¯Q\bar{V}_Q.

The attention mechanism is applied to extract the question-related information from the story.

The machine then goes through the story by the attention mechanism several times and obtain an answer selection vector V¯Qn\bar{V}_{Q_n}. This answer selection vector is used to evaluate the confidence of each choice, and the choice with the highest score is taken as the output.

All the model parameters are jointly trained with the target where 1 for the correct choice and 0 otherwise.

Question Representation

Fig. 3 (A) The Question Vector Representation and (B) The Attention Mechanism.

Fig.3 (A) shows the procedure of encoding the input question into a vector representation V¯Q\bar{V}_Q.

The input question is a sequence of TT words, w1,w2,,wTw_1,w_2, \cdots, w_T, every word WiW_i represented in 1-of-N encoding.

A bidirectional Gated Recurrent Unit (GRU) network [1]-[3] takes one word from the input question sequentially at a time.

In Fig.3(A), the hidden layer output of the forward GRU (green rectangle) at time index tt is denoted by yf(t)y_f(t), and that of the backward GRU (blue rectangle) is by yb(t)y_b(t). After looking through all the words in the question, the hidden layer output of forward GRU network at the last time index yf(T)y_f(T), and that of backward GRU network at the first time index yb(1)y_b(1), are concantenated to form the question vector representation V¯Q\bar{V}_Q, or V¯Q=[yf(T)yb(1)]\bar{V}_Q=[y_f(T)||y_b(1)].


The symbol [][ \cdot || \cdot ] denotes concatenation of two vectors.


Story Attention Module

Fig.3B shows the attention mechanism which takes the question vector V¯Q\bar{V}_Q obtained in Fig.2A and the story transcriptions as the input to encode the whole story into a story vector representations V¯S\bar{V}_S.

The story transcription is a long word sequence with many sentences, only two sentences are shown, and each sentence is with 4 words for simplicity.

There is a bidirectional GRU in Fig.3(B) encoding the whole story into a story vector representation V¯S\bar{V}_S.

The word vector representation of the tt-th word S¯t\bar{S}_t is constructed by a concatenating the hidden layer outputs of forward and backward GRU networks, which is S¯t=[yf(t)yb(t)]\bar{S}_t=[y_f(t) \Vert y_b(t)].

Then the attention value αt\alpha_t for each time index tt is the cosine similarity between the question vector V¯Q\bar{V}_Q and the word vector representation S¯t\bar{S}_t of each word, αt=S¯tV¯Q\alpha_t = \bar{S}_t \odot \bar{V}_Q.

With attention values αt\alpha_t, there can be two different attention mechanisms, word-level and sentence-level, to encode the whole story into the story vector representations V¯s\bar{V}_s.


The symbol \odot denotes cosine similarity between two vectors.


Word-level Attention

All the attention values α\alpha are normalized into αt\alpha_t' such that they sum to one over the whole story. Then all the word vector S¯t\bar{S}_t from the bidirectional GRU network for every word in the story are weighted with this normalized attention value αt\alpha_t' and sum to give the story vector, i.e., V¯S=tαtS¯t\bar{V}_S=\displaystyle \sum_{t} \alpha_t' \bar{S}_t

Sentence-Level Attention

Sentence-level attention means the model collects the information only att the end of each sentence. Therefore, the normalization is only performed over those words tat the end of the sentences to obtain αt"\alpha_t".

The story vector representation is then V¯S=t=Eosαt×S¯t\bar{V}_S= \displaystyle \sum_{t=\textrm{Eos}} \alpha_t'' \times \bar{S}_t, where only those words at the end of sentences (Eos) contribute to the weighted sum. So V¯S=α4×S¯4+αt×S¯8\bar{V}_S = \alpha_4'' \times \bar{S}_4 + \alpha_t'' \times \bar{S}_8 in the example of fig.2.

Hopping

The overall picture of the proposed model is shown in fig.2, in which fig.3 (A) and (B) of the complete proposed model. In the left of fig.2, the input question is first converted into a question vector V¯Q\bar{V}_Q by module in Fig.3(A). This V¯Q\bar{V}_Q is used to compute the attention values αt\alpha_t to obtain the story vector V¯S1\bar{V}_{S_1} by the module in Fig.3(B). Then V¯Q0\bar{V}_{Q_0} and V¯S1\bar{V}_{S_1} are summed to form a new question vector V¯Q1\bar{V}_{Q_1}.

This process is called the first hop (hop1) in fig.2.

The output of the first hop V¯Q1\bar{V}_{Q_1} can be used to compute the new attention to obtain a new story vector VS1V_{S_1}.

This can be considered as the machine going over the story again to re-focus the story with a new question vector.

Again, V¯Q\bar{V}_Q and V¯S1\bar{V}_{S_1} are summed to form V¯Q2\bar{V}_{Q_2} (hop2).

After nn hops, the output of the last hop V¯Qn\bar{V}_{Q_n} is used for the answer selection in Answer Selection.

Answer Selection

As in the upper part of fig.1, the same way previously used to encode the question into V¯Q\bar{V}_Q in fig.2(A) is used here to encode four choice into choice vector representation V¯A\bar{V}_A, V¯B,\bar{V}_B, V¯C\bar{V}_C, V¯D\bar{V}_D.

Then the coin similarity between the output of the last hop V¯Qn\bar{V}_{Q_n}, and the choice vectors are computed, and the choice with highest similarity is chosen.


The symbol [][ \cdot || \cdot ] denotes concatenation of two vectors.

The symbol \odot denotes cosine similarity between two vectors.


[0]

B. H. Tseng, S. S. Shen, H. Y. Lee, L. S. Lee, ``Towards machine comprehension of spoken content: Initial TOEFL listening comprehension test by machine," Towards machine comprehension of spoken content: Initial TOEFL listening comprehension test by machine, 2016.

[1] J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, “Empirical evaluation

of gated recurrent neural networks on sequence modeling,”

arXiv preprint arXiv:1412.3555, 2014.

[2] K. Cho, B. van Merrienboer, D. Bahdanau, and Y. Bengio, “On ¨

the properties of neural machine translation: Encoder-decoder approaches,”

arXiv preprint arXiv:1409.1259, 2014.

[3] D. Bahdanau, K. Cho, and Y. Bengio, “Neural machine translation

by jointly learning to align and translate,” arXiv preprint

arXiv:1409.0473, 2014.

results matching ""

    No results matching ""