Skip to content

use flash_attn_with_kvcache for faster inference (#2539) #1321

use flash_attn_with_kvcache for faster inference (#2539)

use flash_attn_with_kvcache for faster inference (#2539) #1321

Triggered via push December 26, 2023 10:08
Status Success
Total duration 4m 57s
Artifacts

push.yml

on: push
Matrix: lint-and-tests
Fit to window
Zoom out
Zoom in

Annotations

2 warnings
build-docs
The following actions uses node12 which is deprecated and will be forced to run on node16: actions/checkout@v2, actions/setup-python@v2. For more info: https://github.blog/changelog/2023-06-13-github-actions-all-actions-will-run-on-node16-instead-of-node12-by-default/
lint-and-tests (3.8)
The following actions uses node12 which is deprecated and will be forced to run on node16: actions/checkout@v2, actions/setup-python@v2. For more info: https://github.blog/changelog/2023-06-13-github-actions-all-actions-will-run-on-node16-instead-of-node12-by-default/