Releases: foundation-model-stack/fms-hf-tuning
Releases · foundation-model-stack/fms-hf-tuning
v1.2.0
Dependency Updates
- Update packaging requirement from less than 23.2 to a max of 24 to less than 23.2 to max of 25
API/Interface Changes
- Add optional save_model_dir where final checkpoint is saved. See https://github.com/foundation-model-stack/fms-hf-tuning/blob/main/README.md#saving-checkpoints-while-training
Full List of What's Changed
- Add config_utils tests by @aluu317 in #262
- bug: On save event added to callback by @seshapad in #256
- feat: All metric handling changes by @seshapad in #263
- feat: Configuration to set logging level for trigger log by @seshapad in #241
- Data custom collator by @Ssukriti in #260
- feat: per process state metric by @HarikrishnanBalagopal in #239
- feat: Add a dockerfile argument to enable aimstack by @dushyantbehl in #261
- Set default value of target_modules to be None in LoraConfig by @willmj in #269
- feat: Support pretokenized by @kmehant in #272
- Update packaging requirement from <24,>=23.2 to >=23.2,<25 by @dependabot in #212
- Enabling tests for prompt tuning by @Abhishek-TAMU in #278
- fix: do not add special tokens for custom tokenizer by @kmehant in #279
- fix: bug where the logger was not being used properly by @HarikrishnanBalagopal in #286
- Add functionality to free disk space from Github Actions by @willmj in #287
- Add unit test to verify target_modules defaults correctly by @willmj in #281
- docs: Add documentation on experiment tracking. by @dushyantbehl in #257
- Ensure additional metadata to trackers don't throw error in happy case. by @dushyantbehl in #290
- fix: multiple runid creation bug with distributed training by @dushyantbehl in #268
- feat: logging control operation by @seshapad in #264
- fix run evaluation to get base model path by @anhuong in #273
- Fix: Removal of transformers logger and addition of python native logger by @Abhishek-TAMU in #270
- feat: Added additional events such as on_step_begin, on_optimizer_step, on_substep_end by @seshapad in #293
- Always update setuptools to latest by @jbusche in #288
- Rename all fixtures with correct .jsonl extension by @willmj in #295
- feat: add save_model_dir flag where final checkpoint saved by @anhuong in #291
- feat: Example log controller yaml with training state by @seshapad in #296
New Contributors
Full Changelog: v1.1.0...v1.2.0
v1.2.0-rc.1
What's Changed
- Add config_utils tests by @aluu317 in #262
- bug: On save event added to callback by @seshapad in #256
- feat: All metric handling changes by @seshapad in #263
- feat: Configuration to set logging level for trigger log by @seshapad in #241
- deps: limit peft deps by @anhuong in #274
- Data custom collator by @Ssukriti in #260
- Revert "limit peft deps until investigate (#274)" by @anhuong in #275
- feat: per process state metric by @HarikrishnanBalagopal in #239
- feat: Add a dockerfile argument to enable aimstack by @dushyantbehl in #261
- Set default value of target_modules to be None in LoraConfig by @willmj in #269
- feat: Support pretokenized by @kmehant in #272
- Update packaging requirement from <24,>=23.2 to >=23.2,<25 by @dependabot in #212
- Enabling tests for prompt tuning by @Abhishek-TAMU in #278
- fix: do not add special tokens for custom tokenizer by @kmehant in #279
- fix: bug where the logger was not being used properly by @HarikrishnanBalagopal in #286
- Add functionality to free disk space from Github Actions by @willmj in #287
- Add unit test to verify target_modules defaults correctly by @willmj in #281
- docs: Add documentation on experiment tracking. by @dushyantbehl in #257
- Ensure additional metadata to trackers don't throw error in happy case. by @dushyantbehl in #290
- fix: multiple runid creation bug with distributed training by @dushyantbehl in #268
- feat: logging control operation by @seshapad in #264
- fix run evaluation to get base model path by @anhuong in #273
- Fix: Removal of transformers logger and addition of python native logger by @Abhishek-TAMU in #270
- FIX: Metrics file epoch indexing starting from 0 by @Abhishek-TAMU in #294
- feat: Added additional events such as on_step_begin, on_optimizer_step, on_substep_end by @seshapad in #293
- Always update setuptools to latest by @jbusche in #288
- Rename all fixtures with correct .jsonl extension by @willmj in #295
- feat: add save_model_dir flag where final checkpoint saved by @anhuong in #291
New Contributors
Full Changelog: v1.1.0-rc.1...v1.2.0-rc.1
v1.1.0
What's Changed
- fix: Added correct link in main readme for the trainer-controller readme by @seshapad in #254
- trainer controller doc updates by @alex-jw-brooks in #244
- docs: fix the instructions for running with LORA by @HarikrishnanBalagopal in #265
- refactor code to preprocess datasets by @Ssukriti in #259
- Replace shutil.copytree() to fix permission error by @olson-ibm in #251
- fix: logic for getting tracker config by @HarikrishnanBalagopal in #267
- fix: remove lm_head for granite with llama arch models by @anhuong in #258
Full Changelog: v1.0.0...v1.1.0
v1.1.0-rc.1
What's Changed
- fix: Added correct link in main readme for the trainer-controller readme by @seshapad in #254
- trainer controller doc updates by @alex-jw-brooks in #244
- docs: fix the instructions for running with LORA by @HarikrishnanBalagopal in #265
- refactor code to preprocess datasets by @Ssukriti in #259
- Replace shutil.copytree() to fix permission error by @olson-ibm in #251
- fix: logic for getting tracker config by @HarikrishnanBalagopal in #267
- fix: remove lm_head for granite with llama arch models by @Ssukriti in #258
Full Changelog: v1.0.0-rc.1...v1.1.0-rc.1
v1.0.0
New major features:
- Trainer Controller and ability to stop training earlier using metrics and thresholds. See https://github.com/foundation-model-stack/fms-hf-tuning?tab=readme-ov-file#trainer-controller-framework
- Experiment Tracker while tuning for more insights (choice of frameworks)
Dependency upgrades:
- Transformers to minimum version 4.42.0 needed for tuning of granite with llama models
API /interface changes:
- Introduction of tokenizer field in ModelArguments https://github.com/foundation-model-stack/fms-hf-tuning/blob/main/tuning/config/configs.py#L51 to allow custom tokenizer while tuning
- Deprecation of tokenizer field from PromptTuningConfig and moved to ModelArguments
Full List of what's Changed:
- minor refactor to allow modular functions by @Ssukriti in #224
- feat: support some metrics being 'None' without stopping training by @HarikrishnanBalagopal in #169
- Support custom embedding layer resizing to the desired multiple by @kmehant in #227
- ReIntroduce Package for FMS Accel by @fabianlim in #223
- feat: History based loss metric by @seshapad in #156
- docs: instructions for using the trainer controller framework by @HarikrishnanBalagopal in #214
- feat: support custom tokenizer by @kmehant in #229
- tests: tracker unit tests by @dushyantbehl in #172
- feat: Aim runid export by @dushyantbehl in #90
- Move default operations and metrics to variables by @alex-jw-brooks in #249
- Update README.md by @Ssukriti in #240
- deps: Update transformers to latest and skip broken prompt tuning tests by @Abhishek-TAMU in #246
- deps: Update transformers lower bound version by @Abhishek-TAMU in #250
- V100rc1 release by @olson-ibm in #252
New Contributors
- @Abhishek-TAMU made their first contribution in #246
- @olson-ibm made their first contribution in #252
Full Changelog: v0.4.0...v1.0.0
v1.0.0-rc.1
What's Changed
- minor refactor to allow modular functions by @Ssukriti in #224
- feat: support some metrics being 'None' without stopping training by @HarikrishnanBalagopal in #169
- Support custom embedding layer resizing to the desired multiple by @kmehant in #227
- ReIntroduce Package for FMS Accel by @fabianlim in #223
- feat: History based loss metric by @seshapad in #156
- docs: instructions for using the trainer controller framework by @HarikrishnanBalagopal in #214
- feat: support custom tokenizer by @kmehant in #229
- tests: tracker unit tests by @dushyantbehl in #172
- feat: Aim runid export by @dushyantbehl in #90
- Move default operations and metrics to variables by @alex-jw-brooks in #249
- Update README.md by @Ssukriti in #240
- deps: Update transformers to latest and skip broken prompt tuning tests by @Abhishek-TAMU in #246
- deps: Update transformers lower bound version by @Abhishek-TAMU in #250
New Contributors
- @Abhishek-TAMU made their first contribution in #246
Full Changelog: v0.4.0-rc.3...v1.0.0-rc.1
v0.4.0
Summary of Changes
- Support for LoRA tuning for llama3 and granite (with GPTBigCode) architectures
- Dependencies versions adjustment
What's Changed
- remove merge model for lora tuned adapters by @anhuong in #197
- Add test coverage by @tedhtchang in #171
- Install Acceleration Framework into Training Script by @fabianlim in #157
- deps: limit dependency ranges by @anhuong in #54
- Delete dependabot.yml by @tedhtchang in #207
- add dependabot.yml by @tedhtchang in #208
- Fix additional callbacks by @VassilisVassiliadis in #199
- Update trl by @alex-jw-brooks in #213
- deps: cap transformers at 4.40.2 by @anhuong in #218
- Formatting consolidation main by @Ssukriti in #216
- Fix PyPi publish error caused by direct url reference by @tedhtchang in #219
Full Changelog: v0.3.0...v0.4.0
v0.4.0-rc.3
What's Changed
- remove merge model for lora tuned adapters by @anhuong in #197
- Add test coverage by @tedhtchang in #171
- Install Acceleration Framework into Training Script by @fabianlim in #157
- deps: limit dependency ranges by @anhuong in #54
- Delete dependabot.yml by @tedhtchang in #207
- add dependabot.yml by @tedhtchang in #208
- Fix additional callbacks by @VassilisVassiliadis in #199
- Update trl by @alex-jw-brooks in #213
- deps: cap transformers at 4.40.2 by @anhuong in #218
- Formatting consolidation main by @Ssukriti in #216
- Fix PyPi publish error caused by direct url reference by @tedhtchang in #219
Full Changelog: v0.3.0...v0.4.0-rc.3
v0.4.0-rc.2
Summary of Changes
- Support for LoRA tuning for llama3 and granite (with GPTBigCode) architectures
- Various dependencies versions adjustment
What's Changed
- remove merge model for lora tuned adapters by @anhuong in #197
- Add test coverage by @tedhtchang in #171
- Install Acceleration Framework into Training Script by @fabianlim in #157
- deps: limit dependency ranges by @anhuong in #54
- Delete dependabot.yml by @tedhtchang in #207
- add dependabot.yml by @tedhtchang in #208
- Fix additional callbacks by @VassilisVassiliadis in #199
- Update trl by @alex-jw-brooks in #213
- deps: cap transformers at 4.40.2 by @anhuong in #218
Full Changelog: v0.3.0...v0.4.0-rc.2
v0.4.0-rc.1
What's Changed
- remove merge model for lora tuned adapters by @anhuong in #197
- Add test coverage by @tedhtchang in #171
- Install Acceleration Framework into Training Script by @fabianlim in #157
- deps: limit dependency ranges by @anhuong in #54
- Delete dependabot.yml by @tedhtchang in #207
- add dependabot.yml by @tedhtchang in #208
- Fix additional callbacks by @VassilisVassiliadis in #199
- Update trl by @alex-jw-brooks in #213
Full Changelog: v0.3.0...v0.4.0-rc.1