You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have fine-tuned the bitnet_b1_58-large (https://huggingface.co/1bitLLM/bitnet_b1_58-large) on the Alpaca Instruction Tuning dataset. After conversion, the f32.gguf model is giving proper results. But the i2_s.gguf is just outputting random tokens. Hopefully, the conversion process is correct because the FP32 model is giving correct results. Do I need to manage something or am I missing something when converting custom fine-tuned models?
Following are some results that I am getting using the i2_s model:
### Instruction:
Write about the following topic.
### Input:
Deep Learning
### Response:
Deep Learning ath swe shortNC rev rest throwiseë co /**ab symbols symbolay groundë class strikingast '''rob conjug Search shadow rep lath shadow a'ewewunnwise shadow rep ground ground ground ground ground ground ground throwiserobosesrob whatever shadow by ground ground ground groundew style ground ground ground ground ground groundbody whom rang ground ground ground ground ground ground ground ground ground ground groundew rang groundewoi control rest ground groundew rangiz shadow houredaburgeda a ground ground ground ground ground ground ground ground ground ground ground foodilë shell contactellite reception’ew swearation pro work shadow icon' ritane rangage
It should have been similar to this (f32.gguf model).
### Instruction:
Write about the following topic.
### Input:
Deep Learning
### Response:
Deep Learning is a technique used by computers to learn complex patterns, data and patterns in large amounts of data. It involves using a combination of techniques such as machine learning and deep learning, which can help learn complex patterns and identify patterns in large datasets
Is there an issue with the tokenizer, or something else? Any help is appreciated.
The text was updated successfully, but these errors were encountered:
I have fine-tuned the bitnet_b1_58-large (https://huggingface.co/1bitLLM/bitnet_b1_58-large) on the Alpaca Instruction Tuning dataset. After conversion, the
f32.gguf
model is giving proper results. But thei2_s.gguf
is just outputting random tokens. Hopefully, the conversion process is correct because the FP32 model is giving correct results. Do I need to manage something or am I missing something when converting custom fine-tuned models?Following are some results that I am getting using the i2_s model:
It should have been similar to this (
f32.gguf
model).Is there an issue with the tokenizer, or something else? Any help is appreciated.
The text was updated successfully, but these errors were encountered: