Skip to content

Latest commit

 

History

History
67 lines (45 loc) · 1.68 KB

README.md

File metadata and controls

67 lines (45 loc) · 1.68 KB

Annotated-Transformer-KR

The repository that implementated of Transformer with PyTorch. Also some annotation with detail codes.

Getting Started

$ python main.py --help

A quick-start training is also ready for you. You can check basic settings in run-main.sh

$ sh run-main.sh

Performance: Training Multi30k Result

Trained 30 Steps using NVIDIA GTX 1080 ti, Training excution time with validation: 0 h 36 m 36.5517 s

You can run python translate.py to see translate a random sample sentence.

For Example the attention weights(trg-src weight) are something like figure below. You can check all attentions in figs folder.

Source Sentence:
   a young lassie looking dog is in the snow .
Target Sentence:
   ein junger hund , der aussieht wie lassie , im schnee .
Predicted Sentence:
   ein junger hund schaut im schnee . 
Google Translated Sentence:
   ein junger lassie aussehender hund ist im schnee.

Showing the 6-th layer of trg-src sentence attention with 8 heads.

dec_enc_attns-6

Requirements

python >= 3.6
pytorch >= 1.0.0
torchtext
numpy

TODO

  1. Train bigger datas and make a demo server
  2. Beam Search
  3. Calculate BLEU Scores for Translation Task

references

I checked a lot of references. Please visit them and learn it!