You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
model: deepseek/deepseek-r1-distill-qwen-14b # put your embedding model name here, if it is not openai embedding model, should be /<model_name>
alias: default
api_base: https://api.deepseek.com/v1 # change this according to your embedding model
timeout: 120
type: document_store
provider: qdrant
location: http://qdrant:6333
embedding_model_dim: 3072 # put your embedding model dimension here
timeout: 120
recreate_index: true
please change the llm and embedder names to the ones you want to use
the format of llm and embedder should be .<model_name> such as litellm_llm.gpt-4o-2024-08-06
settings:
column_indexing_batch_size: 50
table_retrieval_size: 10
table_column_retrieval_size: 100
allow_using_db_schemas_without_pruning: false # if you want to use db schemas without pruning, set this to true. It will be faster
allow_intent_classification: true
allow_sql_generation_reasoning: true
query_cache_maxsize: 1000
query_cache_ttl: 3600
langfuse_host: https://cloud.langfuse.com
langfuse_enable: true
logging_level: DEBUG
development: true
historical_question_retrieval_similarity_threshold: 0.9
sql_pairs_similarity_threshold: 0.7
sql_pairs_retrieval_max_size: 10
instructions_similarity_threshold: 0.7
instructions_top_k: 10
The text was updated successfully, but these errors were encountered:
I am using the deepseek model. When using the demo data from the official website, the following error is reported:
Please help me find out where the problem is in the configuration file.
-----config.yaml:-----------
type: llm
provider: litellm_llm
models:
put DEEPSEEK_API_KEY=<your_api_key> in ~/.wrenai/.env
model: deepseek/deepseek-reasoner
alias: default
timeout: 120
kwargs:
n: 1
temperature: 0
response_format:
type: text
model: deepseek/deepseek-chat
timeout: 120
kwargs:
n: 1
temperature: 0
response_format:
type: text
model: deepseek/deepseek-coder
timeout: 120
kwargs:
n: 1
temperature: 0
response_format:
type: json_object
type: embedder
provider: litellm_embedder
models:
define OPENAI_API_KEY=<api_key> in ~/.wrenai/.env if you are using openai embedding model
please refer to LiteLLM documentation for more details: https://docs.litellm.ai/docs/providers
alias: default
api_base: https://api.deepseek.com/v1 # change this according to your embedding model
timeout: 120
type: engine
provider: wren_ui
endpoint: http://wren-ui:3000
type: engine
provider: wren_ibis
endpoint: http://wren-ibis:8000
type: document_store
provider: qdrant
location: http://qdrant:6333
embedding_model_dim: 3072 # put your embedding model dimension here
timeout: 120
recreate_index: true
please change the llm and embedder names to the ones you want to use
the format of llm and embedder should be .<model_name> such as litellm_llm.gpt-4o-2024-08-06
the pipes may be not the latest version, please refer to the latest version: https://raw.githubusercontent.com/canner/WrenAI/<WRENAI_VERSION_NUMBER>/docker/config.example.yaml
type: pipeline
pipes:
embedder: litellm_embedder.default
document_store: qdrant
embedder: litellm_embedder.default
document_store: qdrant
embedder: litellm_embedder.default
document_store: qdrant
llm: litellm_llm.default
embedder: litellm_embedder.default
document_store: qdrant
embedder: litellm_embedder.default
document_store: qdrant
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
llm: litellm_llm.default
embedder: litellm_embedder.default
document_store: qdrant
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
llm: litellm_llm.default
llm: litellm_llm.default
embedder: litellm_embedder.default
document_store: qdrant
llm: litellm_llm.default
document_store: qdrant
embedder: litellm_embedder.default
document_store: qdrant
embedder: litellm_embedder.default
llm: litellm_llm.default
llm: litellm_llm.default
engine: wren_ui
llm: litellm_llm.default
llm: litellm_llm.default
llm: litellm_llm.default
llm: litellm_llm.default
engine: wren_ui
embedder: litellm_embedder.default
document_store: qdrant
embedder: litellm_embedder.default
document_store: qdrant
engine: wren_ibis
document_store: qdrant
document_store: qdrant
settings:
column_indexing_batch_size: 50
table_retrieval_size: 10
table_column_retrieval_size: 100
allow_using_db_schemas_without_pruning: false # if you want to use db schemas without pruning, set this to true. It will be faster
allow_intent_classification: true
allow_sql_generation_reasoning: true
query_cache_maxsize: 1000
query_cache_ttl: 3600
langfuse_host: https://cloud.langfuse.com
langfuse_enable: true
logging_level: DEBUG
development: true
historical_question_retrieval_similarity_threshold: 0.9
sql_pairs_similarity_threshold: 0.7
sql_pairs_retrieval_max_size: 10
instructions_similarity_threshold: 0.7
instructions_top_k: 10
The text was updated successfully, but these errors were encountered: