XTuner Release V0.1.22
What's Changed
- [Refactor] fix internlm2 dispatch by @HIT-cwh in #779
- Fix zero3 compatibility issue for DPO by @Johnson-Wang in #781
- [Fix] Fix map_fn in custom_dataset/sft by @fanqiNO1 in #785
- [Fix] fix configs by @HIT-cwh in #783
- [Docs] DPO and Reward Model documents by @RangiLyu in #751
- Support internlm2.5 by @HIT-cwh in #803
- [Bugs] fix dispatch bugs when model not in LOWEST_TRANSFORMERS_VERSION by @HIT-cwh in #802
- [Docs] fix benchmark table by @HIT-cwh in #801
- [Feature] support output without loss in openai_map_fn by @HIT-cwh in #816
- [Docs] fix typos in sp docs by @HIT-cwh in #821
- [Feature] Support the DatasetInfoHook of DPO training by @xu-song in #787
- [Enhance]: Fix sequence parallel memory bottleneck in DPO & ORPO by @RangiLyu in #830
- [Fix] Fix typo by @bychen7 in #795
- [Fix] fix initialization of ref_llm for full param dpo training with zero-3 by @xu-song in #778
- [Bugs] Fix attn mask by @HIT-cwh in #852
- fix lint by @HIT-cwh in #854
- [Bugs] Fix dispatch attn bug by @HIT-cwh in #829
- [Docs]: update readme and DPO en docs by @RangiLyu in #853
- Added minicpm config file to support sft、qlora、lora、dpo by @LDLINGLINGLING in #847
- fix lint by @HIT-cwh in #856
- bump version to 0.1.22 by @HIT-cwh in #855
New Contributors
- @Johnson-Wang made their first contribution in #781
- @xu-song made their first contribution in #787
- @bychen7 made their first contribution in #795
- @LDLINGLINGLING made their first contribution in #847
Full Changelog: v0.1.21...v0.1.22