@@ -924,8 +924,8 @@ def get_num_layers_by_block_type(
924924 layers_block_type_value = getattr (self .hf_config ,
925925 "layers_block_type" , None )
926926 if layers_block_type_value is None :
927- raise ValueError ("The model is an hybrid without a"
928- "layers_block_type in the hf_config,"
927+ raise ValueError ("The model is an hybrid without a "
928+ "layers_block_type in the hf_config, "
929929 "cannot determine the num of "
930930 f"{ block_type .value } layers" )
931931
@@ -2516,7 +2516,7 @@ def _get_and_verify_dtype(
25162516
25172517 if current_platform .is_hpu () and config_dtype == torch .float16 :
25182518 logger .info (
2519- "For HPU, we cast models to bfloat16 instead of"
2519+ "For HPU, we cast models to bfloat16 instead of "
25202520 "using float16 by default. Please specify `dtype` if you "
25212521 "want to use float16." )
25222522 torch_dtype = torch .bfloat16
@@ -2732,7 +2732,7 @@ def __post_init__(self):
27322732 backend = self .guided_decoding_backend ).backend_name
27332733 if backend not in valid_guided_backends :
27342734 raise ValueError (f"Invalid guided_decoding_backend '{ backend } ,"
2735- f"must be one of { valid_guided_backends } " )
2735+ f" must be one of { valid_guided_backends } " )
27362736
27372737
27382738@dataclass
@@ -3008,7 +3008,7 @@ def uuid(self):
30083008 def model_post_init (self , __context : Any ) -> None :
30093009 if not self .enable_reshape and self .enable_fusion :
30103010 logger .warning_once (
3011- "Fusion enabled but reshape elimination disabled."
3011+ "Fusion enabled but reshape elimination disabled. "
30123012 "RMSNorm + quant (fp8) fusion might not work" )
30133013
30143014 pass_config : PassConfig = Field (default_factory = PassConfig )
@@ -3563,7 +3563,7 @@ def set_current_vllm_config(vllm_config: VllmConfig, check_compile=False):
35633563 logger .warning (
35643564 "`torch.compile` is turned on, but the model %s"
35653565 " does not support it. Please open an issue on GitHub"
3566- "if you want it to be supported." ,
3566+ " if you want it to be supported." ,
35673567 vllm_config .model_config .model )
35683568 _current_vllm_config = old_vllm_config
35693569
0 commit comments