Skip to content

Commit

Permalink
Fix handling of responses without prompt fields
Browse files Browse the repository at this point in the history
  • Loading branch information
njhill authored and dtrifiro committed Oct 7, 2024
1 parent 54c600d commit 33fbd5f
Showing 1 changed file with 14 additions and 13 deletions.
27 changes: 14 additions & 13 deletions src/vllm_tgis_adapter/grpc/grpc_server.py
Original file line number Diff line number Diff line change
Expand Up @@ -394,7 +394,7 @@ async def is_cancelled() -> bool:
last_engine_response = result
# In chunked prefill case it's possible that there will be
# multiple prompt-only outputs
if first_response is not None or (
if first_response is None or (
result.prompt_token_ids and not generated_token_count
):
if first_response is None:
Expand Down Expand Up @@ -475,19 +475,20 @@ def _convert_input_details(
response: GenerationResponse,
tokenizer: PreTrainedTokenizer,
) -> GenerationResponse:
response.input_token_count = len(result.prompt_token_ids)
if resp_options.input_tokens:
self._convert_tokens(
result.prompt_token_ids,
result.prompt_logprobs,
include_logprobs=resp_options.token_logprobs,
include_ranks=resp_options.token_ranks,
top_n_tokens=resp_options.top_n_tokens,
tokenizer=tokenizer,
token_infos=response.input_tokens,
)
if result.prompt_token_ids:
response.input_token_count = len(result.prompt_token_ids)
if resp_options.input_tokens:
self._convert_tokens(
result.prompt_token_ids,
result.prompt_logprobs,
include_logprobs=resp_options.token_logprobs,
include_ranks=resp_options.token_ranks,
top_n_tokens=resp_options.top_n_tokens,
tokenizer=tokenizer,
token_infos=response.input_tokens,
)

if resp_options.input_text:
if resp_options.input_text and result.prompt:
response.text = (
result.prompt if not response.text else result.prompt + response.text
)
Expand Down

0 comments on commit 33fbd5f

Please sign in to comment.