Skip to content

Building an Encoder-Decoder Transformer Model from scratch in PyTorch (Training a BART Style Transformer)

Notifications You must be signed in to change notification settings

milanimcgraw/Encoder-Decoder-BART-Style-Transformer-Model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 

Repository files navigation

🤖 Creating and Training a Encoder-Decoder Style Model (from Attention is All You Need) From Scratch

In this repo, we'll be working through an example of how we can create, and then train, the original Transformer from the Attention is All You Need paper.

⚙️The colab link to the code is found and (will also be included in this repo) here.

🫂The video walkthrough can be found here.

⚙️The Build Process

Build the Model

Build the major components of an encoder/decoder style transformer network from scratch using PyTorch.

Train the Model

Train our new network on a toy dataset to showcase how the training loop works and how we pass data through our network.

About

Building an Encoder-Decoder Transformer Model from scratch in PyTorch (Training a BART Style Transformer)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published