refactor: rename docker-compose.gpu.yaml to compose.yaml
Simplified compose file naming to follow Docker Compose best practices: - Renamed docker-compose.gpu.yaml to compose.yaml - Updated all references in documentation files (README.md, DEPLOYMENT.md, GPU_DEPLOYMENT_LOG.md, RUNPOD_TEMPLATE.md) - Updated references in scripts (prepare-template.sh) This change enables simpler command syntax: - Before: docker compose -f docker-compose.gpu.yaml up -d orchestrator - After: docker compose up -d orchestrator Generated with Claude Code (https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
@@ -94,7 +94,7 @@ ssh gpu-pivoine
|
||||
cd /workspace/ai/
|
||||
|
||||
# Start orchestrator (always running)
|
||||
docker compose -f docker-compose.gpu.yaml up -d orchestrator
|
||||
docker compose -f compose.yaml up -d orchestrator
|
||||
|
||||
# Orchestrator will automatically manage model services as needed
|
||||
```
|
||||
@@ -217,7 +217,7 @@ models:
|
||||
### Step 2: Add Docker Service
|
||||
|
||||
```yaml
|
||||
# Add to ai/docker-compose.gpu.yaml
|
||||
# Add to ai/compose.yaml
|
||||
services:
|
||||
vllm-llama:
|
||||
build: ./vllm
|
||||
@@ -245,7 +245,7 @@ services:
|
||||
```bash
|
||||
ssh gpu-pivoine
|
||||
cd /workspace/ai/
|
||||
docker compose -f docker-compose.gpu.yaml restart orchestrator
|
||||
docker compose -f compose.yaml restart orchestrator
|
||||
```
|
||||
|
||||
**That's it!** The orchestrator automatically detects the new model.
|
||||
@@ -256,13 +256,13 @@ docker compose -f docker-compose.gpu.yaml restart orchestrator
|
||||
|
||||
```bash
|
||||
# Start orchestrator
|
||||
docker compose -f docker-compose.gpu.yaml up -d orchestrator
|
||||
docker compose -f compose.yaml up -d orchestrator
|
||||
|
||||
# View orchestrator logs
|
||||
docker logs -f ai_orchestrator
|
||||
|
||||
# Restart orchestrator
|
||||
docker compose -f docker-compose.gpu.yaml restart orchestrator
|
||||
docker compose -f compose.yaml restart orchestrator
|
||||
|
||||
# Check active model
|
||||
curl http://100.100.108.13:9000/health
|
||||
@@ -287,10 +287,10 @@ curl http://100.100.108.13:9000/health | jq '.current_model'
|
||||
|
||||
```bash
|
||||
# Manually start a specific model (bypassing orchestrator)
|
||||
docker compose -f docker-compose.gpu.yaml --profile text up -d vllm-qwen
|
||||
docker compose -f compose.yaml --profile text up -d vllm-qwen
|
||||
|
||||
# Stop a model
|
||||
docker compose -f docker-compose.gpu.yaml stop vllm-qwen
|
||||
docker compose -f compose.yaml stop vllm-qwen
|
||||
|
||||
# View model logs
|
||||
docker logs -f ai_vllm-qwen_1
|
||||
@@ -337,10 +337,10 @@ docker logs -f ai_musicgen_1
|
||||
docker logs ai_orchestrator
|
||||
|
||||
# Check if model service exists
|
||||
docker compose -f docker-compose.gpu.yaml config | grep -A 10 "vllm-qwen"
|
||||
docker compose -f compose.yaml config | grep -A 10 "vllm-qwen"
|
||||
|
||||
# Manually test model service
|
||||
docker compose -f docker-compose.gpu.yaml --profile text up -d vllm-qwen
|
||||
docker compose -f compose.yaml --profile text up -d vllm-qwen
|
||||
curl http://localhost:8001/health
|
||||
```
|
||||
|
||||
@@ -354,8 +354,8 @@ ls -l /var/run/docker.sock
|
||||
sudo systemctl restart docker
|
||||
|
||||
# Rebuild orchestrator
|
||||
docker compose -f docker-compose.gpu.yaml build orchestrator
|
||||
docker compose -f docker-compose.gpu.yaml up -d orchestrator
|
||||
docker compose -f compose.yaml build orchestrator
|
||||
docker compose -f compose.yaml up -d orchestrator
|
||||
```
|
||||
|
||||
### Model Switching Too Slow
|
||||
@@ -376,7 +376,7 @@ docker run --rm -it --gpus all \
|
||||
|
||||
```
|
||||
ai/
|
||||
├── docker-compose.gpu.yaml # Main orchestration file
|
||||
├── compose.yaml # Main orchestration file
|
||||
├── .env.example # Environment template
|
||||
├── README.md # This file
|
||||
│
|
||||
|
||||
Reference in New Issue
Block a user