What's new? |
About TF 2.0 autograph warning message - 컴퓨터 |
a = \
1 # multiple line assignments may cause warning messages.
a = 1 # after I modified all the lines into this format then the warning messages have gone.
1 # multiple line assignments may cause warning messages.
a = 1 # after I modified all the lines into this format then the warning messages have gone.
written time : 2020-01-11 19:47:12.0
what is attention penalty..? - 컴퓨터 |
Linhao Dong, Shuang Xu, Bo Xu, Speech-Transformer: A No-Recurrence Sequence-to-Sequence Model for Speech Recognition, ICASSP 2018
Attention Penalty (from Speech Transformer paper):
In addition, we encouraged the model attending to closer positions by adding a bigger penalty on the attention weights of more distant position-pairs.
There is no more specific description about attention penalty.
This is my imagination, adding negative value for non-diagonal elements on scaled_attention_logits except for the first multi-head attention in decoders.
I have no concrete idea about the attention penalty the authors explained.
Attention Penalty (from Speech Transformer paper):
In addition, we encouraged the model attending to closer positions by adding a bigger penalty on the attention weights of more distant position-pairs.
There is no more specific description about attention penalty.
This is my imagination, adding negative value for non-diagonal elements on scaled_attention_logits except for the first multi-head attention in decoders.
I have no concrete idea about the attention penalty the authors explained.
written time : 2020-01-08 22:11:33.0
델리. - 여행 |
이번 긴(?) 여정의 종착지.
written time : 2019-12-20 05:01:35.0