Skip to content

Latest commit

 

History

History
83 lines (62 loc) · 2.71 KB

README.md

File metadata and controls

83 lines (62 loc) · 2.71 KB

NeuralGPU

Code for the Neural GPU model described in http://arxiv.org/abs/1511.08228. The extended version was described in https://arxiv.org/abs/1610.08613.

Requirements:

  • TensorFlow (see tensorflow.org for how to install)

The model can be trained on the following algorithmic tasks:

  • sort - Sort a symbol list
  • kvsort - Sort symbol keys in dictionary
  • id - Return the same symbol list
  • rev - Reverse a symbol list
  • rev2 - Reverse a symbol dictionary by key
  • incr - Add one to a symbol value
  • add - Long decimal addition
  • left - First symbol in list
  • right - Last symbol in list
  • left-shift - Left shift a symbol list
  • right-shift - Right shift a symbol list
  • bmul - Long binary multiplication
  • mul - Long decimal multiplication
  • dup - Duplicate a symbol list with padding
  • badd - Long binary addition
  • qadd - Long quaternary addition
  • search - Search for symbol key in dictionary

It can also be trained on the WMT English-French translation task:

  • wmt - WMT English-French translation (data will be downloaded)

The value range for symbols are defined by the vocab_size flag. In particular, the values are in the range vocab_size - 1. So if you set --vocab_size=16 (the default) then --problem=rev will be reversing lists of 15 symbols, and --problem=id will be identity on a list of up to 15 symbols.

To train the model on the binary multiplication task run:

python neural_gpu_trainer.py --problem=bmul

This trains the Extended Neural GPU, to train the original model run:

python neural_gpu_trainer.py --problem=bmul --beam_size=0

While training, interim / checkpoint model parameters will be written to /tmp/neural_gpu/.

Once the amount of error gets down to what you're comfortable with, hit Ctrl-C to stop the training process. The latest model parameters will be in /tmp/neural_gpu/neural_gpu.ckpt-<step> and used on any subsequent run.

To evaluate a trained model on how well it decodes run:

python neural_gpu_trainer.py --problem=bmul --mode=1

To interact with a model (experimental, see code) run:

python neural_gpu_trainer.py --problem=bmul --mode=2

To train on WMT data, set a larger --nmaps and --vocab_size and avoid curriculum:

python neural_gpu_trainer.py --problem=wmt --vocab_size=32768 --nmaps=256
  --vec_size=256 --curriculum_seq=1.0 --max_length=60 --data_dir ~/wmt

With less memory, try lower batch size, e.g. --batch_size=4. With more GPUs in your system, there will be a batch on every GPU so you can run larger models. For example, --batch_size=4 --num_gpus=4 --nmaps=512 --vec_size=512 will run a large model (512-size) on 4 GPUs, with effective batches of 4*4=16.

Maintained by Lukasz Kaiser (lukaszkaiser)