You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
pyvene library was designed for model interpretability, not for some production use case which requires training and inference efficiency. pyreft is different. It will have some practical use cases, and require all those production-ready training and inference efficiency.
This ticket may require multiple PRs, including changes in pyvene:
Support multigpu training
Support data parallel
Support model parallel
Support deepspeed at all stage, including gradient checkpoint, model sharding, gpu/cpu offloading
Integrate with accelerate
The text was updated successfully, but these errors were encountered:
I fixed the previous issue in pyvene but now encounter this odd issue depending on number for GPUs
compute_loss isn't called by any pyreft/pyvene code but rather huggingface Trainer
output using single gpu:
Debug Info: torch.Size([4, 4, 1]) cuda:0
Intervening...
output using two gpus:
Debug Info: torch.Size([4, 8, 1]) cuda:0
Intervening...
output using three gpus:
Debug Info: torch.Size([4, 12, 1]) cuda:0
Intervening...
Descriptions:
pyvene
library was designed for model interpretability, not for some production use case which requires training and inference efficiency.pyreft
is different. It will have some practical use cases, and require all those production-ready training and inference efficiency.This ticket may require multiple PRs, including changes in
pyvene
:accelerate
The text was updated successfully, but these errors were encountered: