From b9beef283dac76b4868f13560caeba635f9e98b4 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sebastian=20Kr=C3=BCger?= Date: Thu, 27 Nov 2025 01:24:05 +0100 Subject: [PATCH] fix: remove vllm embedding --- services/vllm/config_llama.yaml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/services/vllm/config_llama.yaml b/services/vllm/config_llama.yaml index f1d4422..280343b 100644 --- a/services/vllm/config_llama.yaml +++ b/services/vllm/config_llama.yaml @@ -3,6 +3,6 @@ host: "0.0.0.0" port: 8001 uvicorn-log-level: "info" gpu-memory-utilization: 0.95 -max-model-len: 32768 +max-model-len: 20480 dtype: auto enforce-eager: false