You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[Bugfix] vLLM should check TorchInductor config for compile cache enablement status
Summary:
vLLM should not assume compile cache is enabled when VLLM_DISABLE_COMPILE_CACHE=0. Users may use TORCHINDUCTOR_FORCE_DISABLE_CACHES=1 to disable compile cache at PyTorch level, effectively making it impossible for vLLM compile cache to function.
Signed-off-by: Yanan Cao <gmagogsfm@gmail.com>
0 commit comments