You can use BigDL-LLM to run any Huggingface Transformer models with INT4 optimizations on either servers or laptops. This directory contains example scripts to help you quickly get started using BigDL-LLM to run some popular open-source models in the community. Each model has its own dedicated folder, where you can find detailed instructions on how to install and run it.
Model | Example |
---|---|
LLaMA | link |
LLaMA 2 | link |
MPT | link |
Falcon | link |
ChatGLM | link |
ChatGLM2 | link |
MOSS | link |
Baichuan | link |
Baichuan2 | link |
Dolly-v1 | link |
Dolly-v2 | link |
RedPajama | link |
Phoenix | link |
StarCoder | link |
InternLM | link |
Whisper | link |
Qwen | link |
To run the examples, we recommend using Intel® Xeon® processors (server), or >= 12th Gen Intel® Core™ processor (client).
For OS, BigDL-LLM supports Ubuntu 20.04 or later, CentOS 7 or later, and Windows 10/11.
For better performance, it is recommended to set environment variables on Linux with the help of BigDL-Nano:
pip install bigdl-nano
source bigdl-nano-init