Skip to content

Conversation

hmellor
Copy link
Member

@hmellor hmellor commented Jul 24, 2025

  • Adds docstring to try_get_generation_config explaining why we still might get the generation config from the model when the generation config is set to vLLM
  • Updates the docstring of get_diff_sampling_param to make it more clear too

Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

This pull request refactors the generation config loading by renaming try_get_generation_config to maybe_get_generation_config and improving its logic. The changes are mostly correct and improve clarity. However, I've found a critical issue where a call to the old function name was missed within vllm/config.py, which will lead to a runtime error.

@hmellor
Copy link
Member Author

hmellor commented Jul 24, 2025

/gemini review

@hmellor hmellor changed the title Cleanup-generation-config-loading Only try and load generation config if it will be used Jul 24, 2025
Copy link
Contributor

@gemini-code-assist gemini-code-assist bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Code Review

The code changes introduce the ability to load multiple signed images to a registry using a bulk load feature. However, the refactoring of get_diff_sampling_param has introduced a regression. The logic to apply self.override_generation_config has become unreachable due to early return statements. This breaks the user-facing feature of overriding generation configs via --override-generation-config.

hmellor added 2 commits July 24, 2025 14:57
Signed-off-by: Harry Mellor <[email protected]>
Signed-off-by: Harry Mellor <[email protected]>
Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

Signed-off-by: Harry Mellor <[email protected]>
Copy link

mergify bot commented Jul 28, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @hmellor.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@mergify mergify bot added the needs-rebase label Jul 28, 2025
@mergify mergify bot removed the needs-rebase label Jul 28, 2025
@DarkLight1337 DarkLight1337 enabled auto-merge (squash) July 31, 2025 09:05
Signed-off-by: Harry Mellor <[email protected]>
@github-actions github-actions bot added the ready ONLY add when PR is ready to merge/full CI is needed label Jul 31, 2025
@hmellor hmellor changed the title Only try and load generation config if it will be used Improve documentation of ModelConfig.try_get_generation_config to prevent future confusion Aug 1, 2025
@vllm-bot vllm-bot merged commit 326a1b0 into vllm-project:main Aug 1, 2025
12 of 18 checks passed
@hmellor hmellor deleted the cleanup-generation-config-loading branch August 1, 2025 16:35
npanpaliya pushed a commit to odh-on-pz/vllm-upstream that referenced this pull request Aug 6, 2025
jinzhen-lin pushed a commit to jinzhen-lin/vllm that referenced this pull request Aug 9, 2025
…revent future confusion (vllm-project#21526)

Signed-off-by: Harry Mellor <[email protected]>
Signed-off-by: Jinzhen Lin <[email protected]>
noamgat pushed a commit to noamgat/vllm that referenced this pull request Aug 9, 2025
…revent future confusion (vllm-project#21526)

Signed-off-by: Harry Mellor <[email protected]>
Signed-off-by: Noam Gat <[email protected]>
paulpak58 pushed a commit to paulpak58/vllm that referenced this pull request Aug 13, 2025
…revent future confusion (vllm-project#21526)

Signed-off-by: Harry Mellor <[email protected]>
Signed-off-by: Paul Pak <[email protected]>
diegocastanibm pushed a commit to diegocastanibm/vllm that referenced this pull request Aug 15, 2025
…revent future confusion (vllm-project#21526)

Signed-off-by: Harry Mellor <[email protected]>
Signed-off-by: Diego-Castan <[email protected]>
epwalsh pushed a commit to epwalsh/vllm that referenced this pull request Aug 28, 2025
zhewenl pushed a commit to zhewenl/vllm that referenced this pull request Aug 28, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
llama Related to Llama models ready ONLY add when PR is ready to merge/full CI is needed v1
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants