Skip to content

Conversation

Oliver-ss
Copy link
Contributor

  1. baichuan-7b and baichuan-13b have different architectures in config.json, "BaichuanForCausalLM" for baichuan-13b and "BaiChuanForCausalLM" for baichuan-7b.
  2. The model structures are almost the same except for position embedding. (RoPE for baichuan-7b and Alibi for baichuan-13b.) So most of the code coule be shared.

Copy link
Member

@zhuohan123 zhuohan123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you for your contribution! Just tested this out with 13b model and -tp 4 and it works like a charm!

@zhuohan123 zhuohan123 merged commit 64f23c2 into vllm-project:main Aug 2, 2023
@luohao123
Copy link

@zhuohan123 So this resolved baichuan13b tp issue?

But I think there was another PR merged in before, which one to use?

jikunshang pushed a commit to jikunshang/vllm that referenced this pull request Jan 24, 2025
Signed-off-by: Chendi Xue <[email protected]>
Signed-off-by: Chendi.Xue <[email protected]>
amy-why-3459 pushed a commit to amy-why-3459/vllm that referenced this pull request Sep 15, 2025
### What this PR does / why we need it?
Part of vllm-project#499 
Add qwen2.5-vl test on single npu, v1 engine is excluded because
qwen2.5-vl has some problems with v1 now, at the same time, this test
can also make vllm-project#639 more credible

Signed-off-by: wangli <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants