-
Notifications
You must be signed in to change notification settings - Fork 19
Logs27: Mutual Information
Higepon Taro Minowa edited this page Jun 13, 2018
·
30 revisions
- Training data set
- p_i: "Let's have curry for lunch."
- q_i: "Maybe Coco ichi?"
- p_i+1: "Sounds good."
- Train seq2seq
- X: concat(p_i, q_i)
- Y: p_i+1
- Train seq2seq_backward
- X: p_i+1
- Y: q_i
- RL Training
- Beam Search
- X: concat(p_i, q_i) [batch_size, decoder_length]
- beam_replies: [batch_size, decoder_length, beam_width]
- logits: [batch_size, decoder_length, vocab_size]
- Calc reward
- Get log_prob: [batch_size, decoder_length, beam_width]
- Beam Search
- done Make it possible that beam coexists with infer
- Return infer_logis when beam search
- Get logits for predicted_id
- Have beam_logits.
- Refactoring
- extract attention method.
- Unify the model class?
- Confirm beam_logits is same size as logits and same values.
- for one beam search result get indices
- Fetch logprob from the indices
- reward back? or make it for multiple.
- Build MI model, this is happening when decoding best N results and mutual information.