1919from vllm .tasks import SupportedTask
2020from vllm .utils import make_async
2121from vllm .v1 .outputs import PoolerOutput , SamplerOutput
22- from vllm .worker .worker_base import WorkerBase
22+ from vllm .v1 . worker .worker_base import WorkerBase
2323
2424logger = init_logger (__name__ )
2525
@@ -30,7 +30,7 @@ class ExecutorBase(ABC):
3030 """Base class for all executors.
3131
3232 An executor is responsible for executing the model on one device,
33- or it can be a distributed executor
33+ or it can be a distributed executor
3434 that can execute the model on multiple devices.
3535 """
3636
@@ -83,7 +83,7 @@ def collective_rpc(self,
8383
8484 Returns:
8585 A list containing the results from each worker.
86-
86+
8787 Note:
8888 It is recommended to use this API to only pass control messages,
8989 and set up data-plane communication to pass data.
@@ -100,7 +100,7 @@ def determine_num_available_blocks(self) -> tuple[int, int]:
100100
101101 Returns a tuple `(num_gpu_blocks, num_cpu_blocks)`, where
102102 `num_gpu_blocks` are blocks that are "active" on the device and can be
103- appended to.
103+ appended to.
104104 `num_cpu_blocks` refers to "swapped" blocks in CPU memory and cannot be
105105 appended to.
106106 """
@@ -327,7 +327,7 @@ def _run_workers(
327327 run only in the remote TP workers, not the driver worker.
328328 It will also be run asynchronously and return a list of futures
329329 rather than blocking on the results.
330-
330+
331331 # TODO: simplify and merge with collective_rpc
332332 """
333333 raise NotImplementedError
0 commit comments