f668e06228
feat: add BGE embedding model for concurrent operation with Llama
...
Build and Push RunPod Docker Image / build-and-push (push) Successful in 36s
- Create config_bge.yaml for BAAI/bge-large-en-v1.5 on port 8002
- Reduce Llama VRAM to 70% and context to 16K for concurrent use
- Add BGE service to supervisor with vllm group
🤖 Generated with [Claude Code](https://claude.com/claude-code )
Co-Authored-By: Claude <noreply@anthropic.com >
2025-11-30 19:55:13 +01:00
977b9c0f4f
fix: proxy root path
Build and Push RunPod Docker Image / build-and-push (push) Successful in 15s
2025-11-27 15:36:45 +01:00
6e57b21c78
fix: audiocraft HF_HOME env
Build and Push RunPod Docker Image / build-and-push (push) Successful in 16s
2025-11-27 14:58:09 +01:00
9e4c3f5a9c
fix: supervisor upscale
Build and Push RunPod Docker Image / build-and-push (push) Successful in 14s
2025-11-27 12:27:50 +01:00
44bfa271e3
feat: upscale service
Build and Push RunPod Docker Image / build-and-push (push) Successful in 15s
2025-11-27 12:04:19 +01:00
a08879625c
fix: remove audiocraft autostart
Build and Push RunPod Docker Image / build-and-push (push) Successful in 14s
2025-11-27 10:52:57 +01:00
400e534f17
fix: webdav-sync config
Build and Push RunPod Docker Image / build-and-push (push) Successful in 17s
2025-11-27 09:33:31 +01:00
90fa8a073c
fix: remove vllm embedding
Build and Push RunPod Docker Image / build-and-push (push) Successful in 36s
2025-11-27 01:12:57 +01:00
5e9aa8f25d
fix: HF_HOME for vllm
Build and Push RunPod Docker Image / build-and-push (push) Successful in 14s
2025-11-27 00:20:42 +01:00
5c61ac5c67
Initial commit
Build and Push RunPod Docker Image / build-and-push (push) Successful in 1m28s
2025-11-26 17:15:08 +01:00