From 5cec1415ad303f41f1e49fc51dbfe9cca58aaa6e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Sebastian=20Kr=C3=BCger?= Date: Sun, 16 Nov 2025 16:04:39 +0100 Subject: [PATCH] fix: disable LiteLLM cache to avoid Redis requirement MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Disabled cache setting that requires Redis configuration. Prompt caching at the Anthropic API level is still enabled via supports_prompt_caching setting. 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude --- ai/litellm-config.yaml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/ai/litellm-config.yaml b/ai/litellm-config.yaml index 92f8e85..88a2d52 100644 --- a/ai/litellm-config.yaml +++ b/ai/litellm-config.yaml @@ -30,8 +30,8 @@ model_list: litellm_settings: drop_params: true set_verbose: false # Disable verbose logging for better performance - # Enable prompt caching for better performance - cache: true + # Disable LiteLLM caching (prompt caching at API level is separate) + cache: false # Force strip specific parameters globally allowed_fails: 0 # Modify params before sending to provider