File tree Expand file tree Collapse file tree 3 files changed +15
-13
lines changed Expand file tree Collapse file tree 3 files changed +15
-13
lines changed Original file line number Diff line number Diff line change 7
7
8
8
import pytest
9
9
10
- from vllm .entrypoints .renderer import Renderer
10
+ from vllm .entrypoints .renderer import CompletionRenderer
11
11
12
12
13
13
@dataclass
@@ -41,9 +41,9 @@ def mock_async_tokenizer():
41
41
42
42
@pytest .fixture
43
43
def renderer (mock_model_config , mock_tokenizer ):
44
- return Renderer (model_config = mock_model_config ,
45
- tokenizer = mock_tokenizer ,
46
- async_tokenizer_pool = {})
44
+ return CompletionRenderer (model_config = mock_model_config ,
45
+ tokenizer = mock_tokenizer ,
46
+ async_tokenizer_pool = {})
47
47
48
48
49
49
class TestRenderPrompt :
@@ -161,9 +161,10 @@ async def test_max_length_exceeded(self, renderer):
161
161
162
162
@pytest .mark .asyncio
163
163
async def test_no_tokenizer_for_text (self , mock_model_config ):
164
- renderer_no_tokenizer = Renderer (model_config = mock_model_config ,
165
- tokenizer = None ,
166
- async_tokenizer_pool = {})
164
+ renderer_no_tokenizer = CompletionRenderer (
165
+ model_config = mock_model_config ,
166
+ tokenizer = None ,
167
+ async_tokenizer_pool = {})
167
168
168
169
with pytest .raises (ValueError , match = "No tokenizer available" ):
169
170
await renderer_no_tokenizer .render_prompt (
Original file line number Diff line number Diff line change 60
60
TranslationRequest )
61
61
from vllm .entrypoints .openai .serving_models import OpenAIServingModels
62
62
from vllm .entrypoints .openai .tool_parsers import ToolParser
63
- from vllm .entrypoints .renderer import Renderer
63
+ from vllm .entrypoints .renderer import BaseRenderer , CompletionRenderer
64
64
# yapf: enable
65
65
from vllm .inputs .data import EmbedsPrompt as EngineEmbedsPrompt
66
66
from vllm .inputs .data import TokensPrompt as EngineTokensPrompt
@@ -226,14 +226,15 @@ def __init__(
226
226
AsyncMicrobatchTokenizer ] = {}
227
227
self .log_error_stack = log_error_stack
228
228
229
- def _get_renderer (self , tokenizer : Optional [AnyTokenizer ]) -> Renderer :
229
+ def _get_renderer (self , tokenizer : Optional [AnyTokenizer ]) -> BaseRenderer :
230
230
"""
231
231
Get a Renderer instance with the provided tokenizer.
232
232
Uses shared async tokenizer pool for efficiency.
233
233
"""
234
- return Renderer (model_config = self .model_config ,
235
- tokenizer = tokenizer ,
236
- async_tokenizer_pool = self ._async_tokenizer_pool )
234
+ return CompletionRenderer (
235
+ model_config = self .model_config ,
236
+ tokenizer = tokenizer ,
237
+ async_tokenizer_pool = self ._async_tokenizer_pool )
237
238
238
239
def _get_async_tokenizer (self , tokenizer ) -> AsyncMicrobatchTokenizer :
239
240
"""
Original file line number Diff line number Diff line change @@ -80,7 +80,7 @@ async def render_prompt(
80
80
raise NotImplementedError
81
81
82
82
83
- class Renderer (BaseRenderer ):
83
+ class CompletionRenderer (BaseRenderer ):
84
84
85
85
def __init__ (
86
86
self ,
You can’t perform that action at this time.
0 commit comments