feat: enable CPU offload for CogVideoX model to reduce VRAM usage

- Add enable_sequential_cpu_offload=true to DownloadAndLoadCogVideoModel
- Reduces VRAM from ~20GB to ~12GB at cost of slower inference
- Widget values: [model, precision, quantization, cpu_offload] = ['THUDM/CogVideoX-5b-I2V', 'bf16', 'disabled', true]
- Necessary for 24GB GPU with other services running
This commit is contained in:
2025-11-23 09:47:02 +01:00
parent 404eb6ad0e
commit 385b36b062

View File

@@ -34,7 +34,7 @@
"type": "DownloadAndLoadCogVideoModel",
"pos": [50, 500],
"size": [350, 100],
"widgets_values": ["THUDM/CogVideoX-5b-I2V"],
"widgets_values": ["THUDM/CogVideoX-5b-I2V", "bf16", "disabled", true],
"title": "CogVideoX-5b-I2V Loader",
"flags": {},
"order": 1,