Image released: quay.io/modh/fms-hf-tuning:v2.4.0
Summary of Changes
Acceleration Updates:
- Dataclass args added for accelerated MoE tuning, which can be activated using the new int flag
fast_moe
for the number of expert parallel sharding. - Update function name from
requires_agumentation
torequires_augmentation
. - Note: the lower limit of the
fms-acceleration
library has been increased to 0.6.0.
Data Preprocessor Updates:
- Allows for padding free plugin to be used without response template.
- Allows HF dataset IDs to be passed via the
training_data_path flag
.
Additional Changes:
- Add pad_token to special_tokens_dict when pad_token == eos_token, which improves granite 3.0 + 3.1 quality on the tuning stack.
For full details of changes, see the release notes.
(edited)
Full List of Change
- fix: broken README.md link by @dushyantbehl in #429
- feat: Allow hf dataset id to be passed via training_data_path by @dushyantbehl in #431
- feat: dataclass args for accelerated MoE tuning by @willmj in #390
- feat: allow for padding free plugin to be used without response template by @dushyantbehl in #430
- fix: function name from
requires_agumentation
torequires_augmentation
by @willmj in #434 - fix: Add pad_token to special_tokens_dict when pad_token == eos_token by @Abhishek-TAMU in #436
- chore(deps): upgrade fms-acceleration to >= 0.6 by @willmj in #440
- docs: update granite3 model support by @anhuong in #441
Full Changelog: v2.3.1...v2.4.0