Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Plan to support latest vLLM? #76

Closed
thanhnguyentung95 opened this issue Nov 18, 2024 · 3 comments · Fixed by #77
Closed

Plan to support latest vLLM? #76

thanhnguyentung95 opened this issue Nov 18, 2024 · 3 comments · Fixed by #77

Comments

@thanhnguyentung95
Copy link
Contributor

Due to the following code in vLLM 0.6.2, we cannot serve LoRA adapters on a per-request basis, as it does not support LoRA and Multimodal simultaneously:

        if self.lora_config:
            assert supports_lora(self.model), "Model does not support LoRA"
            assert not supports_multimodal(
                self.model
            ), "To be tested: Multi-modal model with LoRA settings."

Do you have a plan to upgrade the vLLM version for Aria?

@xffxff
Copy link
Collaborator

xffxff commented Nov 18, 2024

@thanhnguyentung95 Yes, we can upgrade vLLM. I'll take a look at this.

@xffxff
Copy link
Collaborator

xffxff commented Nov 18, 2024

Hi @thanhnguyentung95
You can try upgrading vLLM in your local environment and continue development first. There are several indirect dependencies (like PyTorch, transformers), so I'll need some extra time to test the upgrade thoroughly to ensure it doesn't break any existing Aria functionality.

@xffxff
Copy link
Collaborator

xffxff commented Nov 20, 2024

@thanhnguyentung95 I've upgraded vllm to the latest version in #77

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants