Skip to content

Highly commented implementations of Transformers in PyTorch

License

Notifications You must be signed in to change notification settings

warner-benjamin/commented-transformers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Commented Transformers

Highly commented implementations of Transformers in PyTorch for Creating a Transformer From Scratch series:

  1. The Attention Mechanism
  2. The Rest of the Transformer

The layers folder contains implementations for Bidirectional Attention, Causal Attention, and CausalCrossAttention.

The models folder contains single file implementations for GPT-2 and BERT. Both models are compatible with torch.compile(..., fullgraph=True).

About

Highly commented implementations of Transformers in PyTorch

Topics

Resources

License

Stars

Watchers

Forks

Languages