-
Notifications
You must be signed in to change notification settings - Fork 127
padding, softmax, embeddings #13
Copy link
Copy link
Open
Description
Hi,
I have two questions regarding the CAML implementation:
- All the texts in a batch are padded, but the input to the softmax function is not masked. Hence, this implementation also assigns positives attentions to padding tokens, right? Do I miss something here?
- The embedding vector that belongs to the padding tokens does not seem to be fixed to the zero vector. If not, then where is that constraint implemented? (I guess it wouldn't make a difference if 1. was handled differently, i.e. if the attentions for padding vectors would be fixed to 0).
Many thanks!
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels