From e5fb326215bfd824e961e5a1b7bbc5ed5afeb10f Mon Sep 17 00:00:00 2001 From: Woosuk Kwon Date: Sun, 20 Oct 2024 17:57:44 -0700 Subject: [PATCH] mypy --- vllm/v1/engine/llm_engine.py | 2 +- vllm/v1/outputs.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/vllm/v1/engine/llm_engine.py b/vllm/v1/engine/llm_engine.py index 790166a1d1a5..8df941d9df74 100644 --- a/vllm/v1/engine/llm_engine.py +++ b/vllm/v1/engine/llm_engine.py @@ -501,7 +501,7 @@ def start_profile(self) -> None: def stop_profile(self) -> None: pass - def get_tokenizer_group(self): + def get_tokenizer_group(self, *args, **kwargs): return self.tokenizer diff --git a/vllm/v1/outputs.py b/vllm/v1/outputs.py index 96bf04214533..9bb15f62dc14 100644 --- a/vllm/v1/outputs.py +++ b/vllm/v1/outputs.py @@ -62,7 +62,7 @@ class RequestOutput: request_id: str prompt: Optional[str] prompt_token_ids: List[int] - prompt_logprobs: Optional[Dict[int, float]] + prompt_logprobs: Optional[List[Optional[Dict[int, float]]]] outputs: List[CompletionOutput] finished: bool metrics: Optional[RequestMetrics] = None