Skip to content

Commit 50f1520

Browse files
committed
Pre commit
Signed-off-by: sfeng33 <[email protected]>
1 parent 5524c46 commit 50f1520

File tree

2 files changed

+6
-10
lines changed

2 files changed

+6
-10
lines changed

vllm/entrypoints/openai/serving_engine.py

Lines changed: 3 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -65,6 +65,7 @@
6565
from vllm.entrypoints.renderer import BaseRenderer, CompletionRenderer
6666
# yapf: enable
6767
from vllm.inputs.data import EmbedsPrompt as EngineEmbedsPrompt
68+
from vllm.inputs.data import PromptType
6869
from vllm.inputs.data import TokensPrompt as EngineTokensPrompt
6970
from vllm.inputs.parse import parse_and_batch_prompt
7071
from vllm.logger import init_logger
@@ -1109,7 +1110,7 @@ def _load_and_validate_embed(embed: bytes) -> EmbedsPrompt:
11091110
def _log_inputs(
11101111
self,
11111112
request_id: str,
1112-
inputs: Union[RequestPrompt, EngineTokensPrompt],
1113+
inputs: Union[RequestPrompt, PromptType],
11131114
params: Optional[Union[SamplingParams, PoolingParams,
11141115
BeamSearchParams]],
11151116
lora_request: Optional[LoRARequest],
@@ -1121,11 +1122,9 @@ def _log_inputs(
11211122
prompt = inputs
11221123
elif isinstance(inputs, list):
11231124
prompt_token_ids = inputs
1124-
elif "prompt_embeds" in inputs:
1125-
prompt_embeds = inputs.get("prompt_embeds")
11261125
else:
11271126
prompt = getattr(inputs, 'prompt', None)
1128-
prompt_token_ids = inputs["prompt_token_ids"]
1127+
prompt_token_ids = getattr(inputs, 'prompt_token_ids', None)
11291128

11301129
self.request_logger.log_inputs(
11311130
request_id,

vllm/entrypoints/openai/serving_pooling.py

Lines changed: 3 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@
44
import asyncio
55
import base64
66
import time
7-
from collections.abc import AsyncGenerator, Sequence
7+
from collections.abc import AsyncGenerator
88
from typing import Final, Literal, Optional, Union, cast
99

1010
import jinja2
@@ -26,7 +26,7 @@
2626
PoolingRequest, PoolingResponse,
2727
PoolingResponseData, UsageInfo)
2828
# yapf: enable
29-
from vllm.entrypoints.openai.serving_engine import OpenAIServing, RequestPrompt
29+
from vllm.entrypoints.openai.serving_engine import OpenAIServing
3030
from vllm.entrypoints.openai.serving_models import OpenAIServingModels
3131
from vllm.entrypoints.utils import _validate_truncation_size
3232
from vllm.logger import init_logger
@@ -127,14 +127,11 @@ async def create_pooling(
127127

128128
engine_prompts = await self.io_processor.pre_process_async(
129129
prompt=validated_prompt, request_id=request_id)
130-
request_prompts: Sequence[RequestPrompt] = [
131-
""
132-
] * len(engine_prompts)
133130

134131
elif isinstance(request, PoolingChatRequest):
135132
(
136133
_,
137-
request_prompts,
134+
_,
138135
engine_prompts,
139136
) = await self._preprocess_chat(
140137
request,

0 commit comments

Comments
 (0)