Skip to content

Commit 5b5b01e

Browse files
committed
grammar
Signed-off-by: Can-Zhao <[email protected]>
1 parent 41dabe4 commit 5b5b01e

File tree

1 file changed

+6
-4
lines changed

1 file changed

+6
-4
lines changed

generation/maisi/README.md

Lines changed: 6 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -84,11 +84,13 @@ When `autoencoder_sliding_window_infer_size` is equal to or larger than the late
8484

8585

8686
### Training GPU Memory Usage
87-
VAE is trained on patches and thus can be trained with 16G GPU if patch size is set to be small like [64,64,64].
88-
Users can adjust patch size to fit the GPU memory.
89-
For the released model, we first trained the autoencoder with 16G V100 with small patch size [64,64,64], then continued training with 32G V100 with patch size of [128,128,128].
87+
The VAE is trained on patches and can be trained using a 16G GPU if the patch size is set to a small value, such as [64, 64, 64].
88+
Users can adjust the patch size to fit the available GPU memory.
89+
For the released model, we initially trained the autoencoder on a 16G V100 GPU with a small patch size of [64, 64, 64], and then continued training on a 32G V100 GPU with a larger patch size of [128, 128, 128].
90+
91+
The DM and ControlNet are trained on whole images rather than patches.
92+
The GPU memory usage during training depends on the size of the input images.
9093

91-
DM and ControlNet are trained on the whole images instead of patches. The training GPU memory usage depends on the input image size.
9294
| image size | latent size | Peak Memory |
9395
|--------------|:------------- |:-----------:|
9496
| 256x256x128 | 4x64x64x32 | 5G |

0 commit comments

Comments
 (0)