This is the implementation of GPT2 paper in PyTorch. This is just the architecture. It has 162262609 (162M) parameters (not training on my GPU so doesn't matter:) ).
- 
                Notifications
    
You must be signed in to change notification settings  - Fork 0
 
yashghogre/GPT2-Implementation
About
This is the implementation of GPT2 architecture in PyTorch.
Resources
Stars
Watchers
Forks
Releases
No releases published
              Packages 0
        No packages published