Skip to content
This repository has been archived by the owner on Jan 15, 2024. It is now read-only.

[Sparse Attention][Performance] Accelerate the performance of sparse attention + Benchmark #1397

Open
sxjscience opened this issue Oct 21, 2020 · 1 comment
Labels
enhancement New feature or request help wanted Extra attention is needed performance Performance issues

Comments

@sxjscience
Copy link
Member

sxjscience commented Oct 21, 2020

We are having ongoing efforts about supporting sparse attention in GluonNLP: #1395. To better accelerate related kernels, we can compare the performance of these potential solutions, including:

@sxjscience sxjscience added enhancement New feature or request help wanted Extra attention is needed labels Oct 21, 2020
@sxjscience
Copy link
Member Author

@ZiyueHuang Created the issue here about how we may use TVM to accelerate the speed.

@sxjscience sxjscience changed the title [Window Attention][Performance] Use TVM + Ansor to accelerate the kernel of window attention [Sparse Attention][Performance] Accelerate the performance of sparse attention + Benchmark Oct 23, 2020
@sxjscience sxjscience added the performance Performance issues label Oct 24, 2020
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
enhancement New feature or request help wanted Extra attention is needed performance Performance issues
Projects
None yet
Development

No branches or pull requests

1 participant