-
Notifications
You must be signed in to change notification settings - Fork 348
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
verl v0.2.1 & v0.3 release checklist #354
Comments
What can I help about the 'tool calling examples' part? |
Will megatron context parallelism be supported in the future? |
Yes. We will use mcore that supports cp by default. |
@BearBiscuit05 See #344, I outlined the main challenge. I think it should be relatively straightforward if veRL can start using I imagine we can have GRPO-trained reasoners in the future that learns when to use tools as part of their |
I talked to vllm maintainer yesterday. It seems that there should be no blocking if we switch from generate to chat. Do you mind give it a try to call chat using SPMD style offline inference? |
Not very familiar with inference, but I think I’m starting to get the hang of it. Does this mean I need to build a new |
You should be able to replace Case 1: Detokenize the tokenized inputs we use for
|
The second choice would incur significant overhead when tokenizing on-the-fly (typically 2x slowdown in generation, which is basically unacceptable). I guess we will need to seek solution for case 1 |
Got it. I'll give it a try. |
It seems that the context parallelism in the model part has not been implemented yet. Is this function currently available? |
Not right now, but if you check this roadmap, once verl upgrades MCore, cp will be support. |
Is it possible to optimize startup time? I noticed when using veRL, it is significantly slower to launch a job than when using Huggingface TRL |
v0.2.1
log_prob_micro_batch_size
is smaller than world_size, and fix the case when "the evaluation dataset size is not divisible by the world_size" Hangs during vllm rollout, no error message #12 (comment)ray start
Add instructions on how to run verl on multi-node #278v0.3
The text was updated successfully, but these errors were encountered: