-
Notifications
You must be signed in to change notification settings - Fork 50
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Test Llama-7b-model #250
Comments
This needs to be assigned to me @gkumbhat . Results on train time with PEFT RANDOM init : accumulate_steps: 1 accumulate_steps: 16 accumulate_steps: 32 Seems like varying accumulation steps keeping all other parameters constant, does not affect train time |
Quality scores:(editing as I run summarization) for sentiment evaluation, 100 examples from train, 100 from predict F1 micro: 31% which is lower than number we saw with flan-t5-xl, but flan-t5 models are known to perform well for these tasks |
testing fine tuning is blocked on #257 |
Description
As a developer of caikit NLP, I want to test Llama-7b models with prompt tuning and fine tuning techniques to evaluate for quality and performance.
Discussion
a. We want to test Llama 7b models for these tasks:
with PEFT tuning - Random and fine tuning.
b. We want to measure train time with varying accumulate steps
Acceptance Criteria
The text was updated successfully, but these errors were encountered: