Skip to content

Commit a9944aa

Browse files
authored
fix: typos (#18151)
Signed-off-by: omahs <73983677+omahs@users.noreply.github.com>
1 parent a8f5aec commit a9944aa

File tree

10 files changed

+18
-18
lines changed

10 files changed

+18
-18
lines changed

csrc/attention/attention_kernels.cuh

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -172,7 +172,7 @@ __device__ void paged_attention_kernel(
172172

173173
// Load the query to registers.
174174
// Each thread in a thread group has a different part of the query.
175-
// For example, if the the thread group size is 4, then the first thread in
175+
// For example, if the thread group size is 4, then the first thread in
176176
// the group has 0, 4, 8, ... th vectors of the query, and the second thread
177177
// has 1, 5, 9, ... th vectors of the query, and so on. NOTE(woosuk): Because
178178
// q is split from a qkv tensor, it may not be contiguous.
@@ -259,7 +259,7 @@ __device__ void paged_attention_kernel(
259259

260260
// Load a key to registers.
261261
// Each thread in a thread group has a different part of the key.
262-
// For example, if the the thread group size is 4, then the first thread in
262+
// For example, if the thread group size is 4, then the first thread in
263263
// the group has 0, 4, 8, ... th vectors of the key, and the second thread
264264
// has 1, 5, 9, ... th vectors of the key, and so on.
265265
for (int i = 0; i < NUM_TOKENS_PER_THREAD_GROUP; i++) {

examples/offline_inference/chat_with_tools.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -68,7 +68,7 @@ def get_current_weather(city: str, state: str, unit: 'str'):
6868
"partly cloudly, with highs in the 90's.")
6969

7070

71-
tool_funtions = {"get_current_weather": get_current_weather}
71+
tool_functions = {"get_current_weather": get_current_weather}
7272

7373
tools = [{
7474
"type": "function",
@@ -122,7 +122,7 @@ def get_current_weather(city: str, state: str, unit: 'str'):
122122
# above defined function
123123
tool_calls = json.loads(output)
124124
tool_answers = [
125-
tool_funtions[call['name']](**call['arguments']) for call in tool_calls
125+
tool_functions[call['name']](**call['arguments']) for call in tool_calls
126126
]
127127

128128
# append the answer as a tool message and let the LLM give you an answer

tests/lora/test_lora_huggingface.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ def test_load_checkpoints_from_huggingface(lora_fixture_name, request):
3030

3131
lora_path = get_adapter_absolute_path(lora_name)
3232

33-
# lora loading should work for either absolute path and hugggingface id.
33+
# lora loading should work for either absolute path and huggingface id.
3434
peft_helper = PEFTHelper.from_local_dir(lora_path, 4096)
3535
lora_model = LoRAModel.from_local_checkpoint(
3636
lora_path,

tests/model_executor/weight_utils.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -20,11 +20,11 @@ def test_hf_transfer_auto_activation():
2020
try:
2121
# enable hf hub transfer if available
2222
import hf_transfer # type: ignore # noqa
23-
HF_TRANFER_ACTIVE = True
23+
HF_TRANSFER_ACTIVE = True
2424
except ImportError:
25-
HF_TRANFER_ACTIVE = False
25+
HF_TRANSFER_ACTIVE = False
2626
assert (huggingface_hub.constants.HF_HUB_ENABLE_HF_TRANSFER ==
27-
HF_TRANFER_ACTIVE)
27+
HF_TRANSFER_ACTIVE)
2828

2929

3030
def test_download_weights_from_hf():

vllm/config.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -297,7 +297,7 @@ class ModelConfig:
297297
- 1K -> 1024\n
298298
- 25.6k -> 25,600"""
299299
spec_target_max_model_len: Optional[int] = None
300-
"""Specify the the maximum length for spec decoding draft models."""
300+
"""Specify the maximum length for spec decoding draft models."""
301301
quantization: Optional[QuantizationMethods] = None
302302
"""Method used to quantize the weights. If `None`, we first check the
303303
`quantization_config` attribute in the model config file. If that is

vllm/lora/ops/triton_ops/lora_expand_op.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -153,7 +153,7 @@ def _lora_expand(
153153
lora_token_start_loc (torch.Tensor): A cumulative sum of
154154
num_tokens_per_lora. lora_token_start_loc[0] is always 0 so that
155155
lora_token_start_loc[i], along with num_tokens_per_lora[i]
156-
identifies the the region in token_indices_sorted_by_lora_ids that
156+
identifies the region in token_indices_sorted_by_lora_ids that
157157
LoRA lora_ids[i] should process.
158158
lora_ids (torch.Tensor): LoRA ids to process.
159159
no_lora_flag_cpu (torch.Tensor): A CPU tensor of size 1, that indicates

vllm/model_executor/layers/mamba/mamba_mixer2.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -142,7 +142,7 @@ def mamba_v2_sharded_weight_loader(
142142
) -> LoaderFunction:
143143
"""Create a weight loader for mamba v2. This ensures that the projections
144144
are correctly sharded so that they can be split into x, B, C. It also
145-
ensures the the all the groups corresponding to a head shard is placed
145+
ensures that all the groups corresponding to a head shard is placed
146146
together with it.
147147
"""
148148

vllm/model_executor/models/granite_speech.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,7 @@
2121
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
2222
# See the License for the specific language governing permissions and
2323
# limitations under the License.
24-
"""Inference-only IBM Granite speeech model."""
24+
"""Inference-only IBM Granite speech model."""
2525
import math
2626
from collections.abc import Iterable, Mapping
2727
from typing import Optional, TypedDict, Union
@@ -626,7 +626,7 @@ def _build_input_features_mask(
626626
audio_embed_sizes: torch.Tensor,
627627
) -> torch.Tensor:
628628
"""Calculate the input features mask, which will generally be used
629-
to mask the the padded features for all entries in the batch except
629+
to mask the padded features for all entries in the batch except
630630
for those with the most audio features.
631631
632632
Args:

vllm/model_executor/models/phi4mm_audio.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -91,9 +91,9 @@ class ConformerEncoderLayer(nn.Module):
9191
if set to True, use GLULinear module,
9292
otherwise, used GLUPointWiseConv module.
9393
default to False.
94-
attention_innner_dim: int, optional
94+
attention_inner_dim: int, optional
9595
if equal to -1, attention dim for linears k/q/v is
96-
equal to d_model. otherwise attention_innner_dim is used.
96+
equal to d_model. otherwise attention_inner_dim is used.
9797
default -1.
9898
attention_glu_type: str, optional
9999
activation function for glu used in the multihead attention,
@@ -148,7 +148,7 @@ def __init__(
148148
conv_glu_type="sigmoid",
149149
bias_in_glu=True,
150150
linear_glu_in_convm=False,
151-
attention_innner_dim=-1,
151+
attention_inner_dim=-1,
152152
attention_glu_type="swish",
153153
activation_checkpointing="",
154154
export=False,
@@ -169,7 +169,7 @@ def __init__(
169169
n_head,
170170
d_model,
171171
dropout_rate,
172-
attention_innner_dim,
172+
attention_inner_dim,
173173
attention_glu_type,
174174
bias_in_glu,
175175
use_pt_scaled_dot_product_attention=

vllm/v1/request.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@ def __init__(
7272
assert len(self.mm_inputs) == len(self.mm_hashes)
7373

7474
# Read-only views
75-
# Prevent directly appending to the these lists since
75+
# Prevent directly appending to these lists since
7676
# they should also be updated simultaneously.
7777
self.output_token_ids = ConstantList(self._output_token_ids)
7878
self.all_token_ids = ConstantList(self._all_token_ids)

0 commit comments

Comments
 (0)