-
-
Notifications
You must be signed in to change notification settings - Fork 10.1k
[CI] Optimize entrypoints API server tests #23896
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
wow! great job! |
tests/entrypoints/openai/embedding_tests/test_encoder_decoder.py
Outdated
Show resolved
Hide resolved
tests/entrypoints/openai/embedding_tests/test_optional_middleware.py
Outdated
Show resolved
Hide resolved
"--reasoning-parser", "deepseek_r1", | ||
"--enable-auto-tool-choice", | ||
"--tool-call-parser", "hermes", | ||
"--disable-log-stats", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
you can remove --disable-log-stats
and --disable-log-requests
68a2e19
to
1b41751
Compare
This pull request has merge conflicts that must be resolved before it can be |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @csahithi this is great!!
Replaced large models with smaller ones -
hmellor/tiny-random-LlamaForCausalLM
,microsoft/DialoGPT-small
Is the reason for the latter that the former doesn't have a chat template?
If so we can just ask @hmellor to add the llama 3.2 chat template and replace them all with that.
Oh sorry I see that it does already have a chat template. Then I'm curious what's the reason for using microsoft/DialoGPT-small
too?
I know you have ideas for possible further streamlining but in the interests of incremental improvement could we get this merged first?
Could you fix the merge conflicts and we can see what the new CI timings are like after that too.
If anything needs changing about |
4732592
to
e799966
Compare
aac1623
to
fba4775
Compare
CI failures look related |
This pull request has merge conflicts that must be resolved before it can be |
Signed-off-by: Sahithi Chigurupati <[email protected]>
Signed-off-by: Sahithi Chigurupati <[email protected]>
fba4775
to
bedeffd
Compare
Signed-off-by: Sahithi Chigurupati <[email protected]>
bedeffd
to
d43d48b
Compare
|
||
with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: | ||
yield remote_server | ||
MODEL_NAME = "hmellor/tiny-random-LlamaForCausalLM" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could this also be made a fixture so that we don't have to keep anything in sync with conftest?
|
||
with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: | ||
yield remote_server | ||
MODEL_NAME = "hmellor/tiny-random-LlamaForCausalLM" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could this also be made a fixture so that we don't have to keep anything in sync with conftest?
from ...utils import RemoteOpenAIServer | ||
|
||
MODEL_NAME = "Qwen/Qwen2.5-1.5B-Instruct" | ||
MODEL_NAME = "hmellor/tiny-random-LlamaForCausalLM" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could this also be made a fixture so that we don't have to keep anything in sync with conftest?
] | ||
with RemoteOpenAIServer(MODEL_NAME, args) as remote_server: | ||
yield remote_server | ||
MODEL_NAME = "hmellor/tiny-random-LlamaForCausalLM" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could this also be made a fixture so that we don't have to keep anything in sync with conftest?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
To make paths shorter should we rename the following directories:
basic_tests
->basic
correctness
(unchanged)embedding_tests
->embedding
individual_tests
->individual
lora_tests
->lora
multimodal_tests
->multimodal
|
||
# Use a small embeddings model for faster startup and smaller memory footprint. | ||
# Since we are not testing any chat functionality, | ||
# using a chat capable model is overkill. | ||
MODEL_NAME = "intfloat/multilingual-e5-small" | ||
MODEL_NAME = "hmellor/tiny-random-LlamaForCausalLM" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could this also be made a fixture so that we don't have to keep anything in sync with conftest?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The video_server
fixture is not used here, is that intentional?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The vision_server
fixture is not used here, is that intentional?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The vision_server
fixture is not used here, is that intentional?
This pull request has merge conflicts that must be resolved before it can be |
@csahithi is out this week, we'll see if someone else can take this over |
@debroy-rh has offered to work on this |
Purpose
hmellor/tiny-random-LlamaForCausalLM
,microsoft/DialoGPT-small
Test Plan
Test Result
Essential Elements of an Effective PR Description Checklist
supported_models.md
andexamples
for a new model.