Skip to content

Commit 9b3705b

Browse files
committed
lint
Signed-off-by: Chih-Chieh-Yang <7364402+cyang49@users.noreply.github.com>
1 parent 4c672ff commit 9b3705b

File tree

2 files changed

+6
-6
lines changed

2 files changed

+6
-6
lines changed

vllm/model_executor/layers/mamba/ops/ssd_chunk_scan.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -518,7 +518,7 @@ def _chunk_scan_fwd(
518518
# no in this case no point to use initial states
519519
initial_states = None
520520
elif chunk_indices is None and chunk_offsets is None:
521-
# if chunk_indices and chunk_offsets both unset, then derive
521+
# if chunk_indices and chunk_offsets both unset, then derive
522522
# from seq_idx
523523
chunk_indices, chunk_offsets = seq_idx_to_chunk_indices_offsets(
524524
seq_idx, chunk_size)

vllm/model_executor/models/bamba.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,6 @@
11
# SPDX-License-Identifier: Apache-2.0
22
"""Inference-only Bamba model."""
33
# Added by the IBM Team, 2024
4-
import math
54
from typing import Iterable, Optional, Set, Tuple
65

76
import torch
@@ -21,9 +20,9 @@
2120
from vllm.model_executor.layers.logits_processor import LogitsProcessor
2221
from vllm.model_executor.layers.mamba.mamba_mixer2 import (
2322
MambaMixer2, extra_groups_for_head_shards)
24-
from vllm.model_executor.layers.quantization import QuantizationConfig
2523
from vllm.model_executor.layers.mamba.ops.ssd_chunk_scan import (
2624
seq_idx_to_chunk_indices_offsets)
25+
from vllm.model_executor.layers.quantization import QuantizationConfig
2726
from vllm.model_executor.layers.rotary_embedding import get_rope
2827
from vllm.model_executor.layers.sampler import SamplerOutput, get_sampler
2928
from vllm.model_executor.layers.vocab_parallel_embedding import (
@@ -258,6 +257,7 @@ def forward(
258257
"mamba": BambaMixerDecoderLayer
259258
}
260259

260+
261261
class BambaModel(nn.Module):
262262

263263
def __init__(self, *, vllm_config: VllmConfig, prefix: str = ""):
@@ -329,10 +329,10 @@ def forward(
329329
seq_idx[srt:end] = i
330330
seq_idx.unsqueeze_(0)
331331

332-
# compute metadata for chunked prefill.
333-
# actually this is only needed if there are
332+
# compute metadata for chunked prefill.
333+
# actually this is only needed if there are
334334
# initial states, but this is determinable
335-
# only from attention metadata yet
335+
# only from attention metadata yet
336336
# unavailable from the current top-level forward.
337337
# Rather than complicating things to extract said
338338
# metadata, we simply just compute redundently and

0 commit comments

Comments
 (0)