Distributed learning with mpi
Dependencies: OpenMPI
and mpi4py
(v. >= 2.0.0), keras
(v. >= 1.2.0)
The original package was implemented by Dustin Anderson and evolution of optimization from Thong Nguyen were transfered to mpi-learn and mpi-opt for practical purpose. Both eventually merged in this repository.
Test with the MNIST dataset, with keras+tensorflow
git clone https://github.com/vlimant/NNLO.git
cd NNLO
Example with mnist provided in a python file
python3 models/get_mnist.py
mpirun -np 3 --tag-output python3 TrainingDriver.py --model examples/example_mnist.py --loss categorical_crossentropy --epochs 3
mpirun -np 3 --tag-output python3 TrainingDriver.py --model examples/example_mnist_torch.py --loss categorical_crossentropy --epochs 3
Example with the cifar10 with model json
python3 models/BuildModel.py cifar10
python3 models/get_cifar10.py
mpirun -np 3 --tag-output python3 TrainingDriver.py --model cifar10_arch.json --train train_cifar10.list --val test_cifar10.list --loss categorical_crossentropy --epochs 5
Example of training mnist with 2 workers, each with 2 process per Horovod ring
mpirun -np 5 --tag-output python3 TrainingDriver.py --model examples/example_mnist.py --loss categorical_crossentropy --epochs 3 --n-processes 2
Example of training mnist with early stopping
mpirun -np 3 --tag-output python3 TrainingDriver.py --model examples/example_mnist.py --loss categorical_crossentropy --epochs 10000 --early "val_loss,~<,4"
Example of training with a fixed target
mpirun -np 3 --tag-output python3 TrainingDriver.py --model examples/example_mnist.py --loss categorical_crossentropy --epochs 10000 --target-metric "val_acc,>,0.97"
Example of training the LCD GAN training, for 5 epochs, and checkpointing at each epoch (note that the input data is located on private servers, please ask us to get access to it)
python models/get_3d.py
mpirun -tag-output -n 3 python3 MPIGDriver.py dummy.json train_3d.list test_1_3d.list --loss dummy --epochs 5 --master-gpu --features-name X --labels-name y --tf --mode easgd --worker-optimizer rmsprop --checkpoint ganGP --checkpoint-int 1
And restoring from the previous state
mpirun -tag-output -n 3 python3 MPIGDriver.py dummy.json train_3d.list test_1_3d.list --loss dummy --epochs 5 --master-gpu --features-name X --labels-name y --tf --mode easgd --worker-optimizer rmsprop --checkpoint ganCP --checkpoint-int 1 --restore ganCP
TrainingDriver.py
will load a keras model of your choice and train it on the input data you provide. The script has three required arguments:
- Path to JSON file specifying the Keras model (your model can be converted to JSON using the model's
to_json()
method) - File containing a list of training data. This should be a simple text file with one input data file per line. By default the script expects data stored in HDF5 format; see below for instructions for handling arbitrary input data.
- File containing a list of validation data. This should be a simple text file with one input data file per line.
See TrainingDriver.py
for supported optional arguments. Run the script via mpirun
or mpiexec
. It automatically detects available NVIDIA GPUs and allocate them among the MPI worker processes.
The provided TrainingDriver.py
script handles the case of a model that is specified in JSON format and training data that is stored in HDF5 files. However, the construction of the model and the loading of input data are easily customized.
Use the ModelBuilder class to specify how your model should be constructed: mpi_learn/train/model.py
To specify your model, create a new class deriving from ModelBuilder and override the build_model()
method. This method should take no arguments and return the Keras model you wish to train.
Provide an instance of ModelBuilder when you construct the MPIManager object (see below). At train time, the build_model
method of the ModelBuilder will be called, constructing the model you specified.
The provided ModelFromJson class is a specialized ModelBuilder that constructs a model from a JSON file (as produced by the to_json()
method of a keras
model). This is usually the easiest way to specify the model architecture.
Use the Data class to specify how batches of training data should be generated: mpi_learn/train/data.py
To specify your training data, create a new class deriving from Data and override the generate_data()
method. The generate_data
method should act as follows:
- yield batches of training data in the form required for training with Keras, i.e. ( [x1, x2, ...], [y1, y2, ...] )
- stop yielding batches and return after one epoch's worth of training data has been generated.
Provide an instance of the Data class when you construct the MPIManager (see below). During training, workers will iterate over the output of the generate_data
method once per epoch, computing the gradient of the loss function on each batch.
Note: generate_data
should not continue to yield training batches forever; rather it should generate one epoch's worth of data before returning.
Use the Algo class to configure the details of the training algorithm: mpi_learn/train/algo.py
Provide an instance of the Algo class when you construct the MPIManager (see below). The Algo constructor takes several arguments that specify aspects of the training process:
optimizer
: supported arguments are'sgd'
,'adadelta'
,'rmsprop'
, and'adam'
. For optimizers that have tunable parameters, please specify the values of those parameters as additional arguments (see mpi_learn/train/optimizer.py for details on the individual optimizers)loss
: loss function, specified as a string, e.g. 'categorical_crossentropy'validate_every
: number of gradient updates to process before performing validation. Set to 0 to disable validation.sync_every
: number of batches for workers to process between gradient updates (default 1)
By default the training is performed using the Downpour SGD algorithm [1].
To instead use Elastic Averaging SGD [2], the following additional settings should be provided:
mode
: specify'easgd'
worker_optimizer
: learning algorithm used by individual worker processeselastic_force
,elastic_lr
,elastic_momentum
: force, learning rate, and momentum parameters for the EASGD algorithm
The algorithm proposed in https://arxiv.org/abs/1805.08469 is available. DOCUMENTATION TO BE COMPLETED
Training is initiated by an instance of the MPIManager class, which initializes each MPI process as a worker or master and prepares the training procedure. The MPIManager is constructed using the following ingredients (see TrainingDriver.py for example usage):
comm
: MPI communicator object, usuallyMPI.COMM_WORLD
data
,algo
,model_builder
: instances of theData
,Algo
, andModelBuilder
classes (see above). These three elements determine most of the details of training.num_epochs
: number of training epochstrain_list
,val_list
: lists of inputs files to use for training and validation. Each MPI process should be able to access any/all of the input files; the MPIManager will split the input files among the available worker processes.callbacks
: list ofkeras
callback objects, to be executed by the master process
Other options are available as well: see mpi_learn/mpi/manager.py
In the default training configuration, one MPI process (process 0) is initialized as 'Master' and all others are initialized as 'Workers'. The Master and each Worker have a copy of the model to be trained. Each Worker has access to a subset of the training data.
During training, a Worker reads one batch of training data and computes the gradient of the loss function on that batch. The Worker sends the gradient to the Master, which uses it to update its model weights. The Master sends the updated model weights to the Worker, which then repeats the process with the next batch of training data.
Description and documentation to be added here.
Example of running hyper-optimization on mnist model
mpirun -np 7 --tag-output python3 OptimizationDriver.py --model examples/example_mnist.py --block-size 3 --epochs 3 --num-iterations 10
Example of running hyper-optimization on mnist model, with 2-fold cross validation
mpirun -np 13 --tag-output python3 OptimizationDriver.py --model examples/example_mnist.py --block-size 6 --epochs 5 --num-iterations 10 --n-fold 2
Example of running hyper-optimization on mnist model, with checkpointing every 2 epochs of the masters. And resuming from the last checkpoint
mpirun -np 7 --tag-output python3 OptimizationDriver.py --model examples/example_mnist.py --block-size 3 --epochs 5 --num-iterations 10 --checkpoint CP --checkpoint-interval 2
mpirun -np 7 --tag-output python3 OptimizationDriver.py --model examples/example_mnist.py --block-size 3 --epochs 5 --num-iterations 10 --checkpoint CP --checkpoint-interval 2 --opt-restore
Training the hyperoptimization of an example GAN model
mpirun -np 7 --tag-output python3 OptimizationDriver.py --example gan --block-size 3 --epochs 3 --num-iterations 10 --checkpoint ganOptCP --checkpoint-int 1
and restarting the optimization from where it stopped
mpirun -np 7 --tag-output python3 OptimizationDriver.py --example gan --block-size 3 --epochs 3 --num-iterations 10 --checkpoint ganOptCP --checkpoint-int 1 --opt-restore
[1] Dean et al., Large Scale Distributed Deep Networks. https://research.google.com/archive/large_deep_networks_nips2012.html.
[2] Zhang et al., Deep Learning with Elastic Averaging SGD. https://arxiv.org/abs/1412.6651