- Remove custom PivoineDiffRhythmRun wrapper node - Add git patch file for ComfyUI_DiffRhythm __init__.py - Patch adds LlamaConfig fix at import time - Add arty script 'fix/diffrhythm-patch' to apply patch - Revert all workflows to use original DiffRhythmRun - Remove startup_patch.py and revert start.sh This approach is cleaner and more maintainable than wrapping the node. The patch directly fixes the tensor dimension mismatch (32 vs 64) in DiffRhythm's rotary position embeddings by ensuring num_attention_heads and num_key_value_heads are properly set based on hidden_size. References: - https://github.com/billwuhao/ComfyUI_DiffRhythm/issues/44 - https://github.com/billwuhao/ComfyUI_DiffRhythm/issues/48
ComfyUI Production Workflows
Comprehensive collection of production-ready ComfyUI workflows for RunPod AI Model Orchestrator.
Overview
This directory contains 20 sophisticated, battle-tested workflows designed for production use with the RunPod orchestrator. Each workflow is optimized for 24GB VRAM and includes API compatibility, error handling, and quality gates.
Directory Structure
workflows/
├── text-to-image/ # Text-to-image generation workflows
├── image-to-image/ # Image-to-image transformation workflows
├── image-to-video/ # Image-to-video animation workflows
├── text-to-music/ # Text-to-music generation workflows
├── upscaling/ # Image upscaling and enhancement workflows
├── advanced/ # Advanced multi-model workflows
├── templates/ # Reusable workflow templates
├── README.md # This file
└── WORKFLOW_STANDARDS.md # Workflow development standards
Workflows by Category
Text-to-Image (4 workflows)
| Workflow | Model | Speed | Quality | Use Case |
|---|---|---|---|---|
flux-schnell-t2i-production-v1.json |
FLUX.1-schnell | Fast (4 steps) | Good | Rapid prototyping, iteration |
flux-dev-t2i-production-v1.json |
FLUX.1-dev | Medium (20-50 steps) | Excellent | High-quality final images |
sdxl-refiner-t2i-production-v1.json |
SDXL + Refiner | Medium (30+20 steps) | Excellent | Detailed, refined outputs |
sd35-large-t2i-production-v1.json |
SD3.5-large | Medium (28 steps) | Excellent | Latest Stable Diffusion |
Image-to-Image (3 workflows)
| Workflow | Technique | Use Case |
|---|---|---|
ipadapter-style-i2i-production-v1.json |
IP-Adapter | Style transfer, composition |
ipadapter-face-i2i-production-v1.json |
IP-Adapter + Face | Portrait generation, face swap |
ipadapter-composition-i2i-production-v1.json |
IP-Adapter Multi | Complex scene composition |
Image-to-Video (3 workflows)
| Workflow | Model | Length | Use Case |
|---|---|---|---|
cogvideox-i2v-production-v1.json |
CogVideoX-5b | 6s @ 8fps | AI-driven video generation |
svd-i2v-production-v1.json |
SVD | 14 frames | Quick animations |
svd-xt-i2v-production-v1.json |
SVD-XT | 25 frames | Extended animations |
Text-to-Music (4 workflows)
| Workflow | Model | Duration | Use Case |
|---|---|---|---|
musicgen-small-t2m-production-v1.json |
MusicGen-small | 30s | Fast generation, low VRAM |
musicgen-medium-t2m-production-v1.json |
MusicGen-medium | 30s | Balanced quality/speed |
musicgen-large-t2m-production-v1.json |
MusicGen-large | 30s | Highest quality |
musicgen-melody-t2m-production-v1.json |
MusicGen-melody | 30s | Melody conditioning |
Upscaling (3 workflows)
| Workflow | Technique | Scale | Use Case |
|---|---|---|---|
ultimate-sd-upscale-production-v1.json |
Ultimate SD | 2x-4x | Professional upscaling with detailing |
simple-upscale-production-v1.json |
Model-based | 2x-4x | Fast, straightforward upscaling |
face-upscale-production-v1.json |
Face-focused | 2x | Portrait enhancement |
Advanced (3 workflows)
| Workflow | Technique | Use Case |
|---|---|---|
controlnet-fusion-production-v1.json |
Multi-ControlNet | Precise composition control |
animatediff-video-production-v1.json |
AnimateDiff | Text-to-video animation |
batch-pipeline-production-v1.json |
Batch processing | Multiple variations |
Quick Start
Using with ComfyUI Web Interface
- Open ComfyUI at
http://localhost:8188 - Click "Load" button
- Navigate to
/workspace/ai/models/comfyui/workflows/ - Select desired workflow category and file
- Adjust parameters as needed
- Click "Queue Prompt"
Using with RunPod Orchestrator API
# Example: FLUX Schnell text-to-image
curl -X POST http://localhost:9000/api/comfyui/generate \
-H "Content-Type: application/json" \
-d '{
"workflow": "text-to-image/flux-schnell-t2i-production-v1.json",
"inputs": {
"prompt": "A serene mountain landscape at sunset",
"seed": 42,
"steps": 4
}
}'
# Example: Image upscaling
curl -X POST http://localhost:9000/api/comfyui/generate \
-H "Content-Type: application/json" \
-d '{
"workflow": "upscaling/ultimate-sd-upscale-production-v1.json",
"inputs": {
"image": "path/to/image.png",
"scale": 2
}
}'
Workflow Features
All production workflows include:
- API Compatibility: Input/output nodes for orchestrator integration
- Error Handling: Validation, fallback nodes, graceful degradation
- Quality Gates: Preview nodes, checkpoints, validation steps
- VRAM Optimization: Model unloading, efficient memory management
- Documentation: Embedded descriptions, parameter guides
- Versioning: Semantic versioning in filenames
Model Requirements
Required Models (Essential)
These models are required by most workflows and are auto-downloaded by Ansible:
- FLUX.1-schnell: Fast text-to-image (17GB)
- FLUX.1-dev: High-quality text-to-image (23GB)
- SDXL Base + Refiner: Stable Diffusion XL (13GB)
- SD3.5-large: Latest Stable Diffusion (16GB)
- CLIP ViT-L/14: Image-text understanding (1.7GB)
Optional Models
- CogVideoX-5b: Text-to-video, image-to-video (9.7GB)
- SVD/SVD-XT: Image-to-video (10GB)
- MusicGen variants: Text-to-music (1.5-3.4GB)
- IP-Adapter: Image conditioning (varies)
- ControlNet models: Precise control (varies)
Check /workspace/ai/COMFYUI_MODELS.md for complete model list.
VRAM Considerations
All workflows are designed for 24GB VRAM with these optimizations:
- Sequential Loading: Only one heavy model loaded at a time
- Model Unloading: Explicit cleanup between stages
- Attention Slicing: Enabled for large models
- VAE Tiling: For high-resolution processing
- Batch Size Limits: Capped at VRAM-safe values
Performance Tips
For Speed
- Use FLUX Schnell (4 steps) or SDXL base (20 steps)
- Lower resolution: 512x512 or 768x768
- Disable refiners and upscalers
- Use
--lowvramflag if needed
For Quality
- Use FLUX Dev (50 steps) or SDXL + Refiner
- Higher resolution: 1024x1024 or higher
- Enable face enhancement (Impact-Pack)
- Use Ultimate SD Upscale for final output
For VRAM Efficiency
- Enable model unloading between stages
- Use VAE tiling for >1024px images
- Process batches sequentially, not in parallel
- Monitor with
nvidia-smiduring generation
Troubleshooting
Out of Memory (OOM) Errors
# Check VRAM usage
nvidia-smi
# Solutions:
# 1. Lower resolution
# 2. Reduce batch size
# 3. Enable model unloading
# 4. Use tiled VAE
# 5. Restart ComfyUI to clear VRAM
supervisorctl restart comfyui
Missing Models
# Check which models are linked
ls -lah /workspace/ComfyUI/models/diffusers/
ls -lah /workspace/ComfyUI/models/clip_vision/
# Re-run Ansible to download missing models
cd /workspace/ai
ansible-playbook playbook.yml --tags comfyui-models-all
# Re-link models
arty run models/link-comfyui
Workflow Load Errors
# Check ComfyUI logs
supervisorctl tail -f comfyui
# Common issues:
# - Missing custom nodes: Check custom_nodes/ directory
# - Node version mismatch: Update ComfyUI and custom nodes
# - Corrupted workflow: Validate JSON syntax
Development
Creating New Workflows
See WORKFLOW_STANDARDS.md for detailed guidelines on creating production-ready workflows.
Quick checklist:
- Use semantic versioning in filename
- Add API input/output nodes
- Include preview and save nodes
- Add error handling and validation
- Optimize for 24GB VRAM
- Document all parameters
- Test with orchestrator API
Testing Workflows
# Manual test via ComfyUI UI
# 1. Load workflow in ComfyUI
# 2. Set test parameters
# 3. Queue prompt
# 4. Verify output quality
# API test via orchestrator
curl -X POST http://localhost:9000/api/comfyui/generate \
-H "Content-Type: application/json" \
-d @test-payload.json
# Batch test multiple workflows
cd /workspace/ai/models/comfyui/workflows
for workflow in text-to-image/*.json; do
echo "Testing $workflow..."
# Add test logic here
done
Contributing
When adding new workflows:
- Follow naming convention:
{category}-{model}-{type}-production-v{version}.json - Place in appropriate category directory
- Update this README with workflow details
- Add to
comfyui_models.yamlif new models are required - Test with both UI and API
- Document any special requirements or setup
Resources
- ComfyUI Documentation: https://github.com/comfyanonymous/ComfyUI
- Custom Nodes Manager: Install via ComfyUI-Manager in UI
- Model Registry:
/workspace/ai/model-orchestrator/models.yaml - Ansible Playbook:
/workspace/ai/playbook.yml - Orchestrator API: http://localhost:9000/docs
License
MIT License - Part of RunPod AI Model Orchestrator
Support
For issues or questions:
- Check ComfyUI logs:
supervisorctl tail -f comfyui - Check orchestrator logs:
supervisorctl tail -f orchestrator - Review
/workspace/ai/CLAUDE.mdfor troubleshooting - Check GPU status:
nvidia-smi