Skip to content

Commit e799966

Browse files
committed
Remove unnecessary flags and tests
Signed-off-by: Sahithi Chigurupati <[email protected]>
1 parent 0d92875 commit e799966

File tree

6 files changed

+4
-67
lines changed

6 files changed

+4
-67
lines changed

tests/entrypoints/openai/basic_tests/conftest.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@
88
BASIC_SERVER_ARGS = [
99
"--dtype", "bfloat16", "--max-model-len", "1024", "--enforce-eager",
1010
"--max-num-seqs", "32", "--gpu-memory-utilization", "0.7",
11-
"--disable-log-stats", "--disable-log-requests",
1211
"--enable-server-load-tracking", "--chat-template",
1312
"{% for message in messages %}{{message['role'] + ': ' \
1413
+ message['content'] + '\\n'}}{% endfor %}", "--enable-auto-tool-choice",

tests/entrypoints/openai/embedding_tests/conftest.py

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -9,8 +9,7 @@
99
UNIVERSAL_EMBEDDING_ARGS = [
1010
"--runner", "pooling", "--dtype", "bfloat16", "--enforce-eager",
1111
"--max-model-len", "512", "--gpu-memory-utilization", "0.7",
12-
"--max-num-seqs", "4", "--disable-log-stats", "--disable-log-requests",
13-
"--chat-template", DUMMY_CHAT_TEMPLATE
12+
"--max-num-seqs", "4", "--chat-template", DUMMY_CHAT_TEMPLATE
1413
]
1514

1615

tests/entrypoints/openai/embedding_tests/test_encoder_decoder.py

Lines changed: 0 additions & 55 deletions
This file was deleted.

tests/entrypoints/openai/embedding_tests/test_optional_middleware.py

Lines changed: 2 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@
1515
# Use a small embeddings model for faster startup and smaller memory footprint.
1616
# Since we are not testing any chat functionality,
1717
# using a chat capable model is overkill.
18-
MODEL_NAME = "intfloat/multilingual-e5-small"
18+
MODEL_NAME = "hmellor/tiny-random-LlamaForCausalLM"
1919

2020

2121
@pytest.fixture(scope="module")
@@ -27,12 +27,7 @@ def server(request: pytest.FixtureRequest, embedding_server):
2727
passed_params = [passed_params]
2828

2929
if passed_params:
30-
args = [
31-
"--runner", "pooling", "--dtype", "bfloat16", "--enforce-eager",
32-
"--max-model-len", "512", "--max-num-seqs", "4",
33-
"--gpu-memory-utilization", "0.7", "--disable-log-stats",
34-
"--disable-log-requests", *passed_params
35-
]
30+
args = ["--enforce-eager", *passed_params]
3631
with RemoteOpenAIServer(MODEL_NAME, args) as custom_server:
3732
yield custom_server
3833
else:

tests/entrypoints/openai/individual_tests/test_metrics.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,6 @@ def default_server_args():
3939
@pytest.fixture(scope="module",
4040
params=[
4141
"",
42-
"--disable-frontend-multiprocessing",
4342
f"--show-hidden-metrics-for-version={PREV_MINOR_VERSION}",
4443
])
4544
def server(default_server_args, request):

tests/entrypoints/openai/lora_tests/test_default_mm_loras.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,8 @@
88
import pytest_asyncio
99
from huggingface_hub import snapshot_download
1010

11+
from ....conftest import AudioTestAssets
1112
from ....utils import RemoteOpenAIServer
12-
from ...conftest import AudioTestAssets
1313

1414
# NOTE - the tests in this module are currently analogous to test_chat, but are
1515
# separated to avoid OOM killing due to module-scoped servers, since we

0 commit comments

Comments
 (0)