The overview of veScale n-D parallelism is as follows:
(*
is under development)
The Auto-Parallelize block takes the untouched Model from the user and Parallel Plan (given by manual effort, prefined for each model type, or automatically generated from Auto-Plan*) and then parallelizes the single-device model into nD Parallelism across a mesh of devices.
veScale's nD Parallelism follows a decoupled design where each D of parallelism is handled by an independent sub-block (e.g., DModule only handles Tensor & Sequence Parallel, without coupling with other Parallel). In contrast to the conventional coupled design that intertwines all parallelism together, such a decoupled nD Parallelism enjoys composability, debuggability, explainability, and extensibility, all of which are of great value for hyper-scale training in production.
Our 4D parallelism (Tensor, Sequence, Data, and ZeRO2) is as follows:
# zero model code change
from <HuggingFace> import <ModelCls>, <ModelArgs>
# create fake model without actual memory usage (optional)
fake_model = deferred_init(<ModelCls>, <ModelArgs>)
# initialize 4D device mesh
mesh = init_device_mesh("cuda", (dp_zero_size, tp_sp_size), mesh_dim_names=["DP_ZERO", "TP_SP"])
# parallelize model in tp & sp
from <PredefinedPlan> import sharding_plan
real_tp_sp_model = parallelize_module(fake_model, mesh["TP_SP"], sharding_plan)
# parallelize model in dp
ddp_model = DDP(real_tp_sp_model, mesh["DP_ZERO"])
# parallelize model with zero2
doptimizer = DistributedOptimizer(torch.optim.AdamW, models=[ddp_model])
# train model as if on a single device
for x in range(dataset):
loss = ddp_model(x)
loss.backward()
doptimizer.step()
doptimizer.zero_grad()
More examples can be found in: <repo>/examples/
.
Coming Soon