Skip to content

Latest commit

 

History

History
21 lines (11 loc) · 1.47 KB

README-THUDM.md

File metadata and controls

21 lines (11 loc) · 1.47 KB

ChatGLM-6B on Akash Network

the original code repo is here

Introduction

ChatGLM-6B is an open bilingual language model based on General Language Model (GLM) framework, with 6.2 billion parameters. With the quantization technique, users can deploy locally on consumer-grade graphics cards (only 6GB of GPU memory is required at the INT4 quantization level).

ChatGLM-6B uses technology similar to ChatGPT, optimized for Chinese QA and dialogue. The model is trained for about 1T tokens of Chinese and English corpus, supplemented by supervised fine-tuning, feedback bootstrap, and reinforcement learning with human feedback. With only about 6.2 billion parameters, the model is able to generate answers that are in line with human preference.

In order to facilitate downstream developers to customize the model for their own application scenarios, we also implements an parameter-efficient tuning method based on P-Tuning v2(Guidelines). Tuning requires at least 7GB of GPU memory at INT4 quantization level.

Try the online demo on Huggingface Spaces.

Web UI

image

Demo Video

afa683e2261e637003064d45681bc8d1_0_1688193006.mp4