An implementation of the “Attention is all you need” paper without extra bells and whistles,
or difficult syntax.
Note: The only extra thing added is Dropout regularization in some layers and option to use GPU.
python -m pip install -r requirements.txt
|Before Training||After 100 Epoch|
English -> German Europarl dataset