-
-
Notifications
You must be signed in to change notification settings - Fork 5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature]: AssertionError: MolmoForCausalLM does not support LoRA yet. #11431
Comments
The LoRA support for this model is a bit complicated. You can refer to this #10022 to get an understanding first. |
I was able to implement a version of MolmoForCausalLM that does not crash with the lora adapter i've trained, but the results that are given back by the served vLLM server are identical to the base model without the Lora adapter. Here a sample request log from the vLLM server.
Can anyone see anything that's clearly off with this? |
please feel free to submit a PR. |
Your current environment
The output of `python collect_env.py`
Model Input Dumps
No response
🐛 Describe the bug
Apparently MolmoForCasualLM does not yet support Lora adapters, yielding an AssertionError on serving:
AssertionError: MolmoForCausalLM does not support LoRA yet.
I trained a Lora adapter with HF Trainer and would like to use it together with vLLM for fast inference. This seems not implemented yet. I tested this by trying to serve Molmo directly via:
vllm serve allenai/Molmo-7B-D-0924 --enable-lora --trust-remote-code --max-num-seqs 6 --tensor-parallel-size 1 --lora-modules test=$LORA_DIR/checkpoint-25
Are there any plans to get this working or is there a guide somewhere how i can enable lora for Molmo myself?
If all works I'd be open to submit a PR but i'd need some guidance.
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: