2525    from  ray .runtime_env  import  RuntimeEnv 
2626    from  ray .util .placement_group  import  PlacementGroup 
2727
28-     from  vllm .executor . executor_base  import  ExecutorBase 
28+     from  vllm .v1 . executor  import  Executor 
2929else :
3030    RuntimeEnv  =  Any 
3131    PlacementGroup  =  Any 
32-     ExecutorBase  =  Any 
32+     Executor  =  Any 
3333
3434logger  =  init_logger (__name__ )
3535
@@ -189,7 +189,7 @@ class ParallelConfig:
189189    """ray distributed model workers placement group.""" 
190190
191191    distributed_executor_backend : (
192-         str  |  DistributedExecutorBackend  |  type [ExecutorBase ] |  None 
192+         str  |  DistributedExecutorBackend  |  type [Executor ] |  None 
193193    ) =  None 
194194    """Backend to use for distributed model 
195195    workers, either "ray" or "mp" (multiprocessing). If the product 
@@ -511,7 +511,7 @@ def __post_init__(self) -> None:
511511            # We use multiprocessing by default if world_size fits on the 
512512            # current node and we aren't in a ray placement group. 
513513
514-             from  vllm .executor  import  ray_utils 
514+             from  vllm .v1 . executor  import  ray_utils 
515515
516516            backend : DistributedExecutorBackend  =  "mp" 
517517            ray_found  =  ray_utils .ray_is_available ()
@@ -553,6 +553,12 @@ def __post_init__(self) -> None:
553553        if  self .distributed_executor_backend  is  None  and  self .world_size  ==  1 :
554554            self .distributed_executor_backend  =  "uni" 
555555
556+         if  self .max_parallel_loading_workers  is  not   None :
557+             logger .warning (
558+                 "max_parallel_loading_workers is currently " 
559+                 "not supported and will be ignored." 
560+             )
561+ 
556562    @property  
557563    def  use_ray (self ) ->  bool :
558564        return  self .distributed_executor_backend  ==  "ray"  or  (
@@ -563,7 +569,7 @@ def use_ray(self) -> bool:
563569    @model_validator (mode = "after" ) 
564570    def  _verify_args (self ) ->  Self :
565571        # Lazy import to avoid circular import 
566-         from  vllm .executor . executor_base  import  ExecutorBase 
572+         from  vllm .v1 . executor  import  Executor 
567573
568574        # Enable batch invariance settings if requested 
569575        if  vllm_is_batch_invariant ():
@@ -574,17 +580,17 @@ def _verify_args(self) -> Self:
574580            and  not  isinstance (self .distributed_executor_backend , str )
575581            and  not  (
576582                isinstance (self .distributed_executor_backend , type )
577-                 and  issubclass (self .distributed_executor_backend , ExecutorBase )
583+                 and  issubclass (self .distributed_executor_backend , Executor )
578584            )
579585        ):
580586            raise  ValueError (
581587                "Unrecognized distributed executor backend " 
582588                f"{ self .distributed_executor_backend }  . Supported " 
583589                "values are 'ray', 'mp' 'uni', 'external_launcher', " 
584-                 " custom ExecutorBase  subclass or its import path." 
590+                 " custom Executor  subclass or its import path." 
585591            )
586592        if  self .use_ray :
587-             from  vllm .executor  import  ray_utils 
593+             from  vllm .v1 . executor  import  ray_utils 
588594
589595            ray_utils .assert_ray_available ()
590596
0 commit comments