@@ -84,7 +84,6 @@ def __init__(
8484
8585 self .maybe_forbid_compile = runner_utils .ForbidCompile (
8686 ) if envs .VLLM_XLA_CHECK_RECOMPILATION else nullcontext ()
87- logger .info ("TPUModelRunner created!" )
8887
8988 def _verify_chunked_prefill_config (self ):
9089 if (self .scheduler_config .max_num_batched_tokens
@@ -106,9 +105,6 @@ def _init_mesh(self) -> None:
106105 sharding_strategy = \
107106 self .vllm_config .additional_config ["sharding" ]["sharding_strategy" ]
108107 except KeyError :
109- logger .warning (
110- f"No sharding strategy passed! Using default of full model parallelism={ len (self .devices )} "
111- )
112108 sharding_strategy = {"tensor_parallelism" : len (self .devices )}
113109
114110 if os .getenv ("NEW_MODEL_DESIGN" , False ):
@@ -120,20 +116,12 @@ def _init_mesh(self) -> None:
120116 try :
121117 dp = sharding_strategy ["data_parallelism" ]
122118 except KeyError :
123- logger .warning (
124- "No data parallelism passed! Using default value of 1" )
125119 dp = 1
126-
127120 try :
128121 tp = sharding_strategy ["tensor_parallelism" ]
129122 except KeyError :
130- logger .warning (
131- f"No tensor parallelism passed! Using default value of { len (self .devices )} "
132- )
133123 tp = len (self .devices )
134124
135- tp = sharding_strategy ["tensor_parallelism" ]
136-
137125 axis_names = ("data" , "model" )
138126 mesh_shape = (dp , tp )
139127
@@ -277,8 +265,6 @@ def initialize_kv_cache(self, kv_cache_config: KVCacheConfig) -> None:
277265 if has_kv_transfer_group ():
278266 get_kv_transfer_group ().register_kv_caches (self .kv_caches )
279267
280- logger .info (jax .lib .xla_bridge .get_backend ().platform_version )
281-
282268 def _precompile_backbone (self ) -> None :
283269 for num_tokens in self .num_tokens_paddings :
284270 input_ids = np .ones ((num_tokens , ), dtype = np .int32 )
0 commit comments