You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The vLLM documentation only reflects the possibility to use Ray for running Distributed Inference and Serving with vLLM, even though the #4539 issue is merged and v0.4.3 is released with the MultiprocessingGPUExecutor feature included as an alternative to Ray for single-node inferencing.
Suggest a potential alternative/fix
Update the documentation to reflect the possibility of using MultiprocessingGPUExecutor as an alternative to Ray for single-node inferencing.
The text was updated successfully, but these errors were encountered:
📚 The doc issue
The vLLM documentation only reflects the possibility to use Ray for running Distributed Inference and Serving with vLLM, even though the #4539 issue is merged and v0.4.3 is released with the MultiprocessingGPUExecutor feature included as an alternative to Ray for single-node inferencing.
Suggest a potential alternative/fix
Update the documentation to reflect the possibility of using MultiprocessingGPUExecutor as an alternative to Ray for single-node inferencing.
The text was updated successfully, but these errors were encountered: