Skip to content

Latest commit

 

History

History
9 lines (7 loc) · 365 Bytes

README.md

File metadata and controls

9 lines (7 loc) · 365 Bytes

Train and Use Distil NLP Models

With Distil models, can train and check whether the models are trained well very quickly

Dataset: "emotion" dataset from HuggingFace Hub.

  1. Fine Tuning DistilBERT for Text Classification
  • Classify texts to emotions from the trained model
  1. Fine Tuning DistilGPT2 for Text Generation
  • Generate texts from the trained model