Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fix flash kernel on sm 70 #1027

Merged
merged 3 commits into from
Jan 26, 2024
Merged

Conversation

grimoire
Copy link
Collaborator

  • flash attention triton kernel on sm 70 bug fixing
  • profile throughput bug fixing

@lvhan028
Copy link
Collaborator

lvhan028 commented Jan 25, 2024

chat failed on v100

double enter to end input >>> hi, what's your name

<|im_start|>system
You are an AI assistant whose name is InternLM (书生·浦语).
- InternLM (书生·浦语) is a conversational language model that is developed by Shanghai AI Laboratory (上海人工智能实验室). It is designed to be helpful, honest, and harmless.
- InternLM (书生·浦语) can understand and communicate fluently in the language chosen by the user such as English and 中文.
<|im_end|>
<|im_start|>user
hi, what's your name<|im_end|>
<|im_start|>assistant
 S1 real

double enter to end input >>>

@lvhan028 lvhan028 self-requested a review January 25, 2024 10:24
@grimoire
Copy link
Collaborator Author

chat failed on v100

can not reproduce with codellama.
Guess the error of internlm is caused by bf16, which is not supported on sm70.

@lvhan028
Copy link
Collaborator

I used internlm2-chat-7b, which is fp16

@grimoire
Copy link
Collaborator Author

Done

@lvhan028 lvhan028 merged commit 77f5f97 into InternLM:main Jan 26, 2024
4 of 6 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants