- Setup up a new conda env and install required packages
# create conda env conda create -n accessory python=3.10 -y conda activate accessory # install packages pip install -r requirements.txt
- This project relies on apex, which needs to be compiled from source. Please follow the official instructions.
- LLaMA2-Accessory is powered by flash-attention for efficient attention computation:
Note that the
pip install flash-attn --no-build-isolation
flash_attn
module is not supported on all types of GPUs. If it is not applicable on your machine, please setUSE_FLASH_ATTENTION
in accessory/configs/global_configs.py toFalse
. Vanilla attention computation will then be used.