Skip to content
Draft
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
541 changes: 541 additions & 0 deletions verl/models/mcore/qat_patch.py

Large diffs are not rendered by default.

6 changes: 6 additions & 0 deletions verl/trainer/config/engine/megatron.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -72,6 +72,12 @@ override_transformer_config:
# Attention backend to use (flash,fused,unfused,local,auto). Defaults to auto in mcore, flash in verl
attention_backend: flash

# # Quantization method. None for no quantization, "nvfp4" for NVFP4 quantization
quantization: null

# Whether to enable Quantization-Aware Training (QAT). Default False.
enable_qat: False
Comment on lines +76 to +79
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

better to use the same configs as in FSDP to avoid confusing the users


override_mcore_model_config: {}

# oc.select: default val for ref.megatron.use_mbridge
Expand Down
1,056 changes: 1,056 additions & 0 deletions verl/utils/modelopt_qat_utils.py

Large diffs are not rendered by default.

Loading