Text Generation
Transformers
PyTorch
chatts
feature-extraction
conversational
custom_code

Use unique model_type "chatts" instead of reusing "qwen2"

#8

This change is required to correctly auto-load the model configuration class and link it to the implementation in vllm library as being implemented at https://github.com/vllm-project/vllm/pull/16852

The auto-loading is based on model_type maps and requires a unique name to load specific classes.

With this change and the vllm change mentioned above, the model can be loaded without any changes by vllm both for offline and online (OpenAI API) inference.

alexanderchemeris changed pull request status to open
bytedance-research org

Thanks for implementing this and ChatTS support in vLLM!

xiezhe24 changed pull request status to merged
Your need to confirm your account before you can post a new comment.

Sign up or log in to comment