fix(ai): litellm config
This commit is contained in:
@@ -41,7 +41,6 @@ model_list:
|
|||||||
stream_timeout: 600
|
stream_timeout: 600
|
||||||
supports_system_messages: true # Llama supports system messages
|
supports_system_messages: true # Llama supports system messages
|
||||||
stream: true # Enable streaming by default
|
stream: true # Enable streaming by default
|
||||||
max_tokens: 4096 # Cap completion tokens to leave room for input
|
|
||||||
|
|
||||||
# Embeddings - BGE Large (Port 8002)
|
# Embeddings - BGE Large (Port 8002)
|
||||||
- model_name: bge-large-en
|
- model_name: bge-large-en
|
||||||
|
|||||||
Reference in New Issue
Block a user