You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
My understanding is: In the first line, the tensor shape changed from (layer_num, target_length, source_length) to (layer_num, target_length), and to (layer_num).
But I don't understand what's the mean of "4", and why use the layer num to calculate the location?
If there is a problem with my understanding, thanks for pointing out.
The text was updated successfully, but these errors were encountered:
4 means the number of heads used to multihead-attention.
If you edit the return value of multihead attention in pytorch, you can get the attention with (layer_num, head_num, target_length, source_length) shape.
Consequently, r and c means n_layers and n_heads.
Hope that this comment be helpful to you.
Hi,
In
prepare_fastspeech.ipynb
file,about
My understanding is: In the first line, the tensor shape changed from (layer_num, target_length, source_length) to (layer_num, target_length), and to (layer_num).
But I don't understand what's the mean of "4", and why use the layer num to calculate the location?
If there is a problem with my understanding, thanks for pointing out.
The text was updated successfully, but these errors were encountered: