Add fine-tuning support for text-generation
Past due by over 1 year
56% complete
Is your feature request related to a problem? Please describe.
Currently we only support Prompt tuning based methods for tuning base models. These work great for larger language models, however, lacks a bit for smaller models in terms of quality. To enable more flexibility and choice for training, we want to enable fine-tuning of certain transformer based…
Is your feature request related to a problem? Please describe.
Currently we only support Prompt tuning based methods for tuning base models. These work great for larger language models, however, lacks a bit for smaller models in terms of quality. To enable more flexibility and choice for training, we want to enable fine-tuning of certain transformer based models using HF Trainer: https://huggingface.co/docs/transformers/main_classes/trainer
Specifics:
- Caikit Task: Text-generation
- Eval task: Summarization
- Test Models: TBD
- Test Datasets: TBD
- GPU Support: Single Node, multi-GPU support
- Training stream data model: https://github.com/caikit/caikit-nlp/blob/main/caikit_nlp/data_model/generation.py#L54
- Data formats (for caikit-runtime):
csv
https://huggingface.co/docs/transformers/main_classes/trainer