You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
in the Bert4Rec_dataset script, I see that trg_mask and src_mask outputs are same. Are they supposed to be same or different. It would be helpful if you could confirm it @vatsalsaglani
The text was updated successfully, but these errors were encountered:
It's been a long time since I worked on this. But as far as I remember here, the transformer treats source and target sequence as one sequence instead of source and target. The masked positions depict the valid positions for attention and the model can learn from the entire sequence.
As this architecture is encoder-only we can use the same masks given we use the same during prediction/inference.
in the Bert4Rec_dataset script, I see that trg_mask and src_mask outputs are same. Are they supposed to be same or different. It would be helpful if you could confirm it @vatsalsaglani
The text was updated successfully, but these errors were encountered: