fix: correct vLLM service port to 8000
- Updated qwen-2.5-7b port from 8001 to 8000 in models.yaml - Matches actual vLLM server default port configuration - Tested and verified: orchestrator successfully loaded model and generated response
This commit is contained in:
@@ -7,7 +7,7 @@ models:
|
|||||||
type: text
|
type: text
|
||||||
framework: vllm
|
framework: vllm
|
||||||
service_script: models/vllm/server.py
|
service_script: models/vllm/server.py
|
||||||
port: 8001
|
port: 8000
|
||||||
vram_gb: 14
|
vram_gb: 14
|
||||||
startup_time_seconds: 120
|
startup_time_seconds: 120
|
||||||
endpoint: /v1/chat/completions
|
endpoint: /v1/chat/completions
|
||||||
|
|||||||
Reference in New Issue
Block a user