You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
| 256x256x128 | >=[64,64,32], not used | 2 | 14G | 57s | 1s |
@@ -57,6 +58,23 @@ To reduce GPU memory usage, we can either increasing `autoencoder_tp_num_splits`
57
58
Increasing `autoencoder_tp_num_splits` has smaller impact on the generated image quality.
58
59
Yet reducing `autoencoder_sliding_window_infer_size` may introduce stitching artifact and has larger impact on the generated image quality.
59
60
61
+
### Training GPU Memory Usage
62
+
VAE is trained on patches and thus can be trained with 16G GPU if patch size is set to be small like [64,64,64].
63
+
Users can adjust patch size to fit the GPU memory.
64
+
For the released model, we first trained the autoencoder with 16G V100 with small patch size [64,64,64], then continued training with 32G V100 with patch size of [128,128,128].
65
+
66
+
DM and ControlNet training GPU memory usage depends on the input image size.
0 commit comments