Seq2Seq
Seq2Seq
Overview
From scratch implementation of Seq2Seq
Key Features
- Attention Mechanism
Technical Details
- Framework: PyTorch
- Dataset: Custom
- Category: Sequential Models
Implementation Details
Trained a Seq2Seq model with the said attention mechanism coded from scratch in Pytorch
Effective Approaches to Attention-based Neural Machine Translation
Sequence to Sequence Learning with Neural Networks
ModelArgs Hyperparameters
Parameter | Value | Description |
---|---|---|
batch_size |
32 | The number of samples processed before the model is updated. |
max_lr |
1e-4 | Maximum learning rate. |
dropout |
0.1 | Dropout. |
epochs |
50 | Epochs |
block_size |
32 | Seq Len |
No of neurons |
128 | No of neurons in an GRU per layer |
hidden_dim |
4*embedding_dims | No of neurons in FFN |
No of neurons |
128 | No of neurons in an GRU per layer |
Frameworks:
Pytorch
Epochs/Steps
Epochs (train) = 50
Val iterations = every epoch
Loss Curves
π View Training Loss Curves
Source Code
π GitHub Repository: Seq2Seq
View the complete implementation, training scripts, and documentation on GitHub.