model_provider = "litellm" model = "anthropic/claude-sonnet-4-5-20250929" # model = "hosted_vllm/meta-llama/Llama-3.1-8B-Instruct" # Context window settings for Llama-3.1-8B (24K context) # model_context_window = 24576 # model_auto_compact_token_limit = 15000 # model_max_output_tokens = 1024 [projects."/home/valknar"] trust_level = "trusted" [projects."/home/valknar/Projects/llmx"] trust_level = "trusted" [projects."/home/valknar/Projects/docker-compose"] trust_level = "trusted" [projects."/home/valknar/Projects/kit-ui"] trust_level = "trusted" [projects."/home/valknar/Projects/image-ui"] trust_level = "trusted" [projects."/home/valknar/bin"] trust_level = "trusted" [projects."/home/valknar/Projects/pivoine.art"] trust_level = "trusted" [mcp_servers.git] # Git operations - Python uvx (no NPM package available) command = "uvx" args = ["mcp-server-git"] enabled = true startup_timeout_sec = 10 [mcp_servers.playwright] # Browser automation - most popular MCP server (826k weekly downloads) command = "npx" args = ["-y", "@playwright/mcp"] enabled = false # Disabled to reduce prompt size startup_timeout_sec = 20 [mcp_servers.brave_search] # Brave Search - requires API key (BRAVE_API_KEY in .env) command = "/home/valknar/.llmx/run-brave-search-mcp.sh" args = [] enabled = true startup_timeout_sec = 15 [mcp_servers.python_runner] # Secure Python sandbox execution (Python uvx, requires Deno) command = "uvx" args = ["mcp-run-python", "stdio"] enabled = true startup_timeout_sec = 15 [mcp_servers.crawl4ai_rag] # Web crawling + RAG with vector search (local Docker stack) # Uses BGE embeddings via remote LiteLLM, local PostgreSQL + pgvector # Start DB first: cd ~/.llmx && docker compose up -d crawl4ai-db crawl4ai-rest command = "/home/valknar/.llmx/run-crawl4ai-mcp.sh" args = [] enabled = true startup_timeout_sec = 60 tool_timeout_sec = 120