[Model][LoRA]LoRA support added for Qwen#9622
[Model][LoRA]LoRA support added for Qwen#9622DarkLight1337 merged 4 commits intovllm-project:mainfrom
Conversation
|
👋 Hi! Thank you for contributing to the vLLM project. Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can do one of these:
🚀 |
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
ccc2f34 to
6462961
Compare
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
DarkLight1337
left a comment
There was a problem hiding this comment.
Looks good, sorry for making you wait!
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: Randall Smith <Randall.Smith@amd.com>
|
Just realized that the Supported Models page hasn't been updated yet. @jeejeelee can you open a new PR to update that page with the new LoRA support? We should also explicitly inherit from |
Okay, handling it now |
Why do we need to do it? |
Easier to find which models support LoRA. |
Get it! |
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: Loc Huynh <jc1da.3011@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: Sumit Dubey <sumit.dubey2@ibm.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com>
Signed-off-by: Jee Jee Li <pandaleefree@gmail.com> Signed-off-by: LeiWang1999 <leiwang1999@outlook.com>
FILL IN THE PR DESCRIPTION HERE
FIX #3458
FIX #9584
Distinguish between Qwen LLM and VL to better support LoRA (similar treatment needed for ChatGLM as well).
Currently set as WIP, the main purpose is to discuss whether this solution(separate LLM and VL) is acceptable , if accepted, I will continue to complete it.ping @ywang96 @DarkLight1337