File tree Expand file tree Collapse file tree 1 file changed +2
-7
lines changed
vllm/model_executor/models Expand file tree Collapse file tree 1 file changed +2
-7
lines changed Original file line number Diff line number Diff line change 4848 DEFAULT_VOCAB_PADDING_SIZE , ParallelLMHead , VocabParallelEmbedding )
4949from vllm .model_executor .model_loader .weight_utils import (
5050 default_weight_loader , maybe_remap_kv_scale_name )
51- from vllm .platforms import current_platform
51+ from vllm .model_executor . sampling_metadata import SamplingMetadata
5252from vllm .sequence import IntermediateTensors
5353
5454from .interfaces import SupportsEagle3 , SupportsLoRA , SupportsPP
@@ -541,12 +541,7 @@ def __init__(self,
541541 self .unpadded_vocab_size ,
542542 config .hidden_size ,
543543 org_num_embeddings = config .vocab_size ,
544- padding_size = (
545- DEFAULT_VOCAB_PADDING_SIZE
546- # We need bigger padding if using lora for kernel
547- # compatibility
548- if not lora_config else
549- current_platform .get_lora_vocab_padding_size ()),
544+ padding_size = DEFAULT_VOCAB_PADDING_SIZE ,
550545 quant_config = quant_config ,
551546 prefix = maybe_prefix (prefix , "lm_head" ),
552547 )
You can’t perform that action at this time.
0 commit comments