Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How can we use Llama2 here? #10

Open
shivprasad94 opened this issue Aug 4, 2023 · 3 comments
Open

How can we use Llama2 here? #10

shivprasad94 opened this issue Aug 4, 2023 · 3 comments

Comments

@shivprasad94
Copy link

I see from the code repo that we are using OpenAI APIs, how can we make this work for Open source models like Llama2?
can someone give me detail on this and what steps I need to follow?

@stefanhgm
Copy link
Contributor

stefanhgm commented Sep 27, 2023

Hello @shivprasad94,

sorry for the late reply and thanks for reaching out!

TabLLM is LLM agnostic, so you can use whatever LLM you want. For instance, to use another HuggingFace model you could create a new json config in TabLLM/t-few/configs (e.g. llama.json) and use the model specifier for the original_model parameter (e.g., "origin_model": "meta-llama/Llama-2-7b").

You can then use this model configuration in the run configuration few-shot-pretrained-100k.sh in line 18 as for model in 'llama'..

Let us know if you need any further help!

@RyanJJP
Copy link

RyanJJP commented May 6, 2024

There seems to be something wrong with t-few when finetune since LLaMA is not an EncoderDecoder model

@stefanhgm
Copy link
Contributor

Hello @RyanJJP,

thanks for this additional comment. You are right, t-few might not work with LLaMA. However, other fine-tuning methods for LLaMA (e.g. QLoRA) should allow a similar functionality. This would require larger changes to the code basis but conceptually it should be similar.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants