Skip to content

Commit 3b352a2

Browse files
authored
Correct capitalisation: VLLM -> vLLM (#14562)
Signed-off-by: Harry Mellor <19981378+hmellor@users.noreply.github.com>
1 parent dea985a commit 3b352a2

File tree

18 files changed

+25
-25
lines changed

18 files changed

+25
-25
lines changed

benchmarks/kernels/benchmark_rmsnorm.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -139,7 +139,7 @@ def calculate_diff(batch_size, seq_len, hidden_size, use_residual=True):
139139

140140
print(f"Naive output={output_naive}")
141141
print(f"FlashInfer output={output_flashinfer}")
142-
print(f"VLLM output={output_vllm}")
142+
print(f"vLLM output={output_vllm}")
143143

144144
if torch.allclose(output_naive, output_flashinfer, atol=1e-2,
145145
rtol=1e-2) and torch.allclose(

docs/source/contributing/vulnerability_management.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ you may contact the following individuals:
3737

3838
## Slack Discussion
3939

40-
You may use the `#security` channel in the [VLLM Slack](https://slack.vllm.ai)
40+
You may use the `#security` channel in the [vLLM Slack](https://slack.vllm.ai)
4141
to discuss security-related topics. However, please do not disclose any
4242
vulnerabilities in this channel. If you need to report a vulnerability, please
4343
use the GitHub security advisory system or contact a VMT member privately.

docs/source/design/v1/metrics.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -509,7 +509,7 @@ cache to complete other requests), we swap kv cache blocks out to CPU
509509
memory. This is also known as "KV cache offloading" and is configured
510510
with `--swap-space` and `--preemption-mode`.
511511

512-
In v0, [VLLM has long supported beam
512+
In v0, [vLLM has long supported beam
513513
search](gh-issue:6226). The
514514
SequenceGroup encapsulated the idea of N Sequences which
515515
all shared the same prompt kv blocks. This enabled KV cache block

examples/offline_inference/disaggregated_prefill_lmcache.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@
55
We will launch 2 vllm instances (GPU 0 for prefill and GPU 1 for decode),
66
and launch an additional LMCache server.
77
KV cache is transferred in the following manner:
8-
VLLM prefill node -> LMCache server -> VLLM decode node.
8+
vLLM prefill node -> LMCache server -> vLLM decode node.
99
1010
Note that `pip install lmcache` is needed to run this example.
1111
Learn more about LMCache in https://github.com/LMCache/LMCache.

tests/tpu/test_quantization_accuracy.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@ def get_model_args(self) -> str:
2525
GSM8KAccuracyTestConfig(
2626
model_name="neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8",
2727
excepted_value=0.76), # no bias
28-
# NOTE(rob): We cannot re-initialize VLLM in the same process for TPU,
28+
# NOTE(rob): We cannot re-initialize vLLM in the same process for TPU,
2929
# so only one of these tests can run in a single call to pytest. As
3030
# a follow up, move this into the LM-EVAL section of the CI.
3131
# GSM8KAccuracyTestConfig(

vllm/attention/selector.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@ def get_env_variable_attn_backend() -> Optional[_Backend]:
5151
# (default behavior if this variable is None)
5252
#
5353
# THIS SELECTION TAKES PRECEDENCE OVER THE
54-
# VLLM ATTENTION BACKEND ENVIRONMENT VARIABLE
54+
# VLLM_ATTENTION_BACKEND ENVIRONMENT VARIABLE
5555
forced_attn_backend: Optional[_Backend] = None
5656

5757

vllm/compilation/backends.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -278,7 +278,7 @@ def call_module(self, target: torch.fx.node.Target,
278278

279279

280280
class VllmBackend:
281-
"""The compilation backend for `torch.compile` with VLLM.
281+
"""The compilation backend for `torch.compile` with vLLM.
282282
It is used for compilation level of `CompilationLevel.PIECEWISE`,
283283
where we customize the compilation.
284284

vllm/compilation/compiler_interface.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@ def initialize_cache(self, cache_dir: str, disable_cache: bool = False):
3131

3232
def compute_hash(self, vllm_config: VllmConfig) -> str:
3333
"""
34-
Gather all the relevant information from the VLLM config,
34+
Gather all the relevant information from the vLLM config,
3535
to compute a hash so that we can cache the compiled model.
3636
3737
See :meth:`VllmConfig.compute_hash` to check what information

vllm/config.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -3572,11 +3572,11 @@ def __str__(self):
35723572
@contextmanager
35733573
def set_current_vllm_config(vllm_config: VllmConfig, check_compile=False):
35743574
"""
3575-
Temporarily set the current VLLM config.
3575+
Temporarily set the current vLLM config.
35763576
Used during model initialization.
3577-
We save the current VLLM config in a global variable,
3577+
We save the current vLLM config in a global variable,
35783578
so that all modules can access it, e.g. custom ops
3579-
can access the VLLM config to determine how to dispatch.
3579+
can access the vLLM config to determine how to dispatch.
35803580
"""
35813581
global _current_vllm_config
35823582
old_vllm_config = _current_vllm_config
@@ -3611,7 +3611,7 @@ def get_current_vllm_config() -> VllmConfig:
36113611
# in ci, usually when we test custom ops/modules directly,
36123612
# we don't set the vllm config. In that case, we set a default
36133613
# config.
3614-
logger.warning("Current VLLM config is not set.")
3614+
logger.warning("Current vLLM config is not set.")
36153615
from vllm.config import VllmConfig
36163616
return VllmConfig()
36173617
return _current_vllm_config

vllm/entrypoints/openai/protocol.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -237,7 +237,7 @@ class ChatCompletionRequest(OpenAIBaseModel):
237237
tool_choice: Optional[Union[Literal["none"], Literal["auto"],
238238
ChatCompletionNamedToolChoiceParam]] = "none"
239239

240-
# NOTE this will be ignored by VLLM -- the model determines the behavior
240+
# NOTE this will be ignored by vLLM -- the model determines the behavior
241241
parallel_tool_calls: Optional[bool] = False
242242
user: Optional[str] = None
243243

0 commit comments

Comments
 (0)