-
-
Notifications
You must be signed in to change notification settings - Fork 7.7k
[Bug]: Cannot register new model #16228
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
This is an only issue with v1 engine. Everything works fine with v0 engine. |
Are you registering the model using the plugin interface as shown here? |
I followed the instructions here https://docs.vllm.ai/en/stable/contributing/model/registration.html#out-of-tree-models. Do I need to also register as a plugin? |
Ok this worked if I add it as an entrypoint plugin. Thanks! |
Hi~ I register model as an entrypoint plugin, and when loading llm I can see it loaded successfully in logs. But I still get the error, do you have any ideas? Thanks. |
Can you provide more details? |
sorry to bother, this is due to the error in my entrypoint plugin. |
Uh oh!
There was an error while loading. Please reload this page.
Your current environment
The output of `python collect_env.py`
🐛 Describe the bug
I am trying to register a new model architecture that is currently not support in vLLM.
But this fails with
Seems that the issue is that when I register my model architecture on the driver side, the
ModelRegistry
dict gets updated. But when loading the model on the worker side, if I print outModelRegistry
in thevllm.model_executor.model_loader.utils.get_model_architecture
function, the registry does not show the new model architecture that I added.I believe the
ModelRegistry
on the driver side has to be serialized and explicitly passed to the workers?Before submitting a new issue...
The text was updated successfully, but these errors were encountered: