This repository has been archived by the owner on Jan 15, 2024. It is now read-only.
[Sparse Attention][Performance] Accelerate the performance of sparse attention + Benchmark #1397
Labels
enhancement
New feature or request
help wanted
Extra attention is needed
performance
Performance issues
We are having ongoing efforts about supporting sparse attention in GluonNLP: #1395. To better accelerate related kernels, we can compare the performance of these potential solutions, including:
We may try out these implementations
The text was updated successfully, but these errors were encountered: