v0.0.1 Generation Quality Improvements
Pre-release
Pre-release
This release improves quality of generation by:
- Changing the defaults sampling settings for LLaMA from greedy to locally typical sampling.
- Updating
drama_llama
andllama.cpp
for BPE tokenizer changes. This will require regenerating any models. See the linked issue for scripts. - Changing OpenAI sampling settings to better suit story generation.
Known Issues:
- This release has some crashes that are fixed in v0.0.2