fix(ai): litellm config
This commit is contained in:
@@ -41,7 +41,6 @@ model_list:
|
||||
stream_timeout: 600
|
||||
supports_system_messages: true # Llama supports system messages
|
||||
stream: true # Enable streaming by default
|
||||
max_tokens: 4096 # Cap completion tokens to leave room for input
|
||||
|
||||
# Embeddings - BGE Large (Port 8002)
|
||||
- model_name: bge-large-en
|
||||
|
||||
Reference in New Issue
Block a user