Skip to content

v2.4.0

Latest
Compare
Choose a tag to compare
@willmj willmj released this 16 Jan 19:43
76bd76d

Image released: quay.io/modh/fms-hf-tuning:v2.4.0

Summary of Changes

Acceleration Updates:

  • Dataclass args added for accelerated MoE tuning, which can be activated using the new int flag fast_moe for the number of expert parallel sharding.
  • Update function name from requires_agumentation to requires_augmentation.
  • Note: the lower limit of the fms-acceleration library has been increased to 0.6.0.

Data Preprocessor Updates:

  • Allows for padding free plugin to be used without response template.
  • Allows HF dataset IDs to be passed via the training_data_path flag.

Additional Changes:

  • Add pad_token to special_tokens_dict when pad_token == eos_token, which improves granite 3.0 + 3.1 quality on the tuning stack.
    For full details of changes, see the release notes.
    (edited)

Full List of Change

  • fix: broken README.md link by @dushyantbehl in #429
  • feat: Allow hf dataset id to be passed via training_data_path by @dushyantbehl in #431
  • feat: dataclass args for accelerated MoE tuning by @willmj in #390
  • feat: allow for padding free plugin to be used without response template by @dushyantbehl in #430
  • fix: function name from requires_agumentation to requires_augmentation by @willmj in #434
  • fix: Add pad_token to special_tokens_dict when pad_token == eos_token by @Abhishek-TAMU in #436
  • chore(deps): upgrade fms-acceleration to >= 0.6 by @willmj in #440
  • docs: update granite3 model support by @anhuong in #441

Full Changelog: v2.3.1...v2.4.0