Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarification Needed: Does chat-model internally use completion-model during chat responses? #4098

Closed
daoutech-sujinlee opened this issue Apr 1, 2025 · 1 comment
Labels
enhancement New feature or request

Comments

@daoutech-sujinlee
Copy link

daoutech-sujinlee commented Apr 1, 2025

I'm currently using Tabby with both a completion-model and a chat-model configured, like this:
tabby serve --model DeepSeekCoder-1.3B --chat-model Qwen2.5-Coder-7B-Instruct

According to the documentation, the completion model and the chat model operate independently — the completion model is for code suggestions in editors, and the chat model is for conversational queries.

However, during my tests, I observed significantly different chat results depending on which completion model was paired with the same chat model. This led me to wonder:

Does Tabby internally utilize the completion-model in the background to assist chat-model responses?

Or are they completely separate as the docs suggest?

It would be great to have an official clarification on how the models interact during chat-based usage, and whether the completion model has any implicit influence on chat output.

In my case, I tested the following combinations:

Image

This makes it seem like the completion model might be influencing the chat result, even though the chat model is the same.

I'd appreciate any clarification from the team — and if this is the intended behavior, it would be helpful to include in the documentation.
To clarify, all of my tests were done via the Tabby chat feature within the IDE (IntelliJ).

@daoutech-sujinlee daoutech-sujinlee added the enhancement New feature or request label Apr 1, 2025
@zwpaper
Copy link
Member

zwpaper commented Apr 2, 2025

Hello @daoutech-sujinlee, Tabby does use the two models independently.

Based on your post, the issue you're experiencing could be related to the small chat model you're using; it's possible to receive different answers upon retrying the questions, even without any changes being made.

@zwpaper zwpaper closed this as completed Apr 7, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants