fix: correct vLLM service port to 8000

- Updated qwen-2.5-7b port from 8001 to 8000 in models.yaml
- Matches actual vLLM server default port configuration
- Tested and verified: orchestrator successfully loaded model and generated response
This commit is contained in:
2025-11-21 16:28:54 +01:00
parent 9a637cc4fc
commit 57b706abe6

View File

@@ -7,7 +7,7 @@ models:
type: text
framework: vllm
service_script: models/vllm/server.py
port: 8001
port: 8000
vram_gb: 14
startup_time_seconds: 120
endpoint: /v1/chat/completions