Researcher Library: Dynamically Stretch or Compress Workload's GPU Allocation¶
The Run:AI Researcher Library is a python library you can add to your deep learning python code. The library contains an elasticity module that allows train workloads to shrink or expand based on the cluster's availability.
Expanding a Workload¶
Expanding a training Job allows your workload to run on more GPUs than the Researcher code was originally written for. This is useful for maximizing the utilization of the cluster as a whole as well as allowing workloads to run faster if idle GPUs exist in the cluster. The extra GPUs will be automatically reclaimed if needed by other, prioritized jobs.
Shrinking a Workload¶
Shrinking a training Job allows your workload to run on a smaller number of GPUs than the Researcher code was originally written for. This is useful for maximizing utilization of the cluster as a whole as well as allowing a Researcher to run a workload, albeit slower than intended, and let it automatically expand when GPUs become available at a later time.
Shrinking a training Job uses an algorithm called Gradient Accumulation. For more information about the algorithm see: https://towardsdatascience.com/what-is-gradient-accumulation-in-deep-learning-ec034122cfa
Install the Run:AI Python library using the following command:
pip install runai
In your python code, if using Keras, add:
If using PyTorch, add:
To initialize the module, you need two parameters:
Maximum GPU batch size - The maximum batch size that your Job can use on a single GPU (in terms of GPU memory). Without Elasticity, running with batch sizes larger than this number will cause a memory overflow. This number will be used by the Run:AI elasticity module for determining whether to use Gradient Accumulation or not.
Global batch size - The desired batch size. Of course, if this number is larger than the Maximum GPU batch size defined above, the model will not fit into a single GPU. The elasticity module will then use Gradient Accumulation and multiple GPUs to run your code.
init() method from the imported module and pass these two arguments. For example, if you are using PyTorch, use the following line:
Where 256 is the global batch size and 64 is the maximum GPU batch size.
Create a Keras model:
model = runai.elastic.keras.models.Model(model)
For PyTorch models, you'll need to wrap your
Optimizer with Gradient Accumulation:
optimizer = runai.ga.torch.optim.Optimizer(optimizer, runai.elastic.steps)
runai.elastic.steps is the value of accumulated steps which is calculated when calling
In addition, you will need to data-parallelise your model. We recommend using the built-in
model = torch.nn.DataParallel(model)
Running a Training Workload¶
Run the training workload by using the "elastic" flag:
- When launching the Job with the runai submit command use
- When launching a Job via YAML code, use the label "elastic" with the value "true"
For additional information on how to run elastic training workloads, see the following quickstart.
- Elasticity currently works with Keras-based or PyTorch-based deep learning code only.
- Any training Job using Run:AI is subject to pause/resume episodes. Elasticity may increase these episodes, making it even more important to make your code resilient. Take care to save checkpoints in your code and have your code resume from the latest checkpoint rather than start from the beginning.
- For additional documentation as well as Python examples see our GitHub repository
- Keras sample code in Github
- Pytorch Sample code in Github