Skip to content

Commit

Permalink
use max autotune no cuda graphs in 10b
Browse files Browse the repository at this point in the history
use max autotune no cuda graphs in 10b
  • Loading branch information
samsja committed Dec 11, 2024
1 parent d814fff commit f3e0fde
Show file tree
Hide file tree
Showing 4 changed files with 4 additions and 0 deletions.
1 change: 1 addition & 0 deletions configs/10B/H100.toml
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ wandb_resume = false
[train]
micro_bs = 1
ac_ckpt = true
torch_compile_mode = "max-autotune-no-cudagraphs"

[optim]
sched_type = "wsd-sqrt"
Expand Down
1 change: 1 addition & 0 deletions configs/10B/H100_cooldown.toml
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ wandb_resume = false
[train]
micro_bs = 1
ac_ckpt = true
torch_compile_mode = "max-autotune-no-cudagraphs"

[optim]
sched_type = "wsd-sqrt"
Expand Down
1 change: 1 addition & 0 deletions configs/10B/H100_simple.toml
Original file line number Diff line number Diff line change
Expand Up @@ -4,6 +4,7 @@ project = "debug_10B_zero_band"
[train]
micro_bs = 1
ac_ckpt = true
torch_compile_mode = "max-autotune-no-cudagraphs"

[optim]
sched_type = "wsd-sqrt"
Expand Down
1 change: 1 addition & 0 deletions configs/150M_short/H100.toml
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ type_model = "llama2"
[train]
micro_bs = 64 # change this base on the gpu
reshard_after_forward = true
torch_compile_mode = "max-autotune-no-cudagraphs"

[optim]
batch_size = 512
Expand Down

0 comments on commit f3e0fde

Please sign in to comment.