Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

在npu上,对大模型用zero3进行全参微调,初始换参数很大,是什么原因造成的? #4272

Open
1 task done
fjw1049 opened this issue Jun 14, 2024 · 0 comments
Labels
npu This problem is related to NPU devices pending This problem is yet to be addressed

Comments

@fjw1049
Copy link

fjw1049 commented Jun 14, 2024

Reminder

  • I have read the README and searched the existing issues.

System Info

training_loss

Reproduction

model_name_or_path: /data1/mixstral

method

stage: sft
do_train: true
finetuning_type: full

ddp

ddp_timeout: 180000000
deepspeed: examples/deepspeed/ds_z3_offload_config.json

dataset

dataset: sft_zh_data,alpaca_zh_demo
template: llama3
cutoff_len: 1024
max_samples: 1000
overwrite_cache: true
preprocessing_num_workers: 16

output

output_dir: saves/full
logging_steps: 30
save_steps: 1000
plot_loss: true
overwrite_output_dir: true

train

per_device_train_batch_size: 1
gradient_accumulation_steps: 2
learning_rate: 1.0e-4
num_train_epochs: 2.0
lr_scheduler_type: cosine
warmup_ratio: 0.1
fp16: true
flash_attn: auto

eval

val_size: 0.1
per_device_eval_batch_size: 1
eval_strategy: steps
eval_steps: 500

Expected behavior

No response

Others

No response

@github-actions github-actions bot added the pending This problem is yet to be addressed label Jun 14, 2024
@hiyouga hiyouga added the npu This problem is related to NPU devices label Jun 19, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
npu This problem is related to NPU devices pending This problem is yet to be addressed
Projects
None yet
Development

No branches or pull requests

2 participants