Skip to content

Latest commit

 

History

History
24 lines (17 loc) · 1.25 KB

README.md

File metadata and controls

24 lines (17 loc) · 1.25 KB

Transformer

A TensorFlow 2.x implementation of the Transformer from Attention Is All You Need (Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, Illia Polosukhin, arxiv, 2017).

This is my attempt at trying to understand and recreate the transformer from the research paper. This is just for my own understanding of the subject and is by no means perfect.

In order to understand and implement the transformer I've taken the help of various tutorials and code guides, which I'll be linking in the resources section.

Requirements

  • tensorflow==2.1.0
  • numpy==1.16.5
  • tensorflow_datasets==3.2.1

How to run

python train.py

Resources