diff --git a/index.html b/index.html index 49df88a..ee6f7b6 100644 --- a/index.html +++ b/index.html @@ -78,7 +78,8 @@
- Please join us at INTERSPEECH 2024 at Kos Island, Greece! + Please check out our poster at + INTERSPEECH 2024 at Kos Island, Greece!
@@ -113,40 +114,36 @@This benchmark demonstrates how our single-step models stack up with previous methods, - most of which mostly require hundreds of generation steps. + most of which requiring hundreds of generation steps. +
+ + ++
Guidance Method | +CFG Weight | +Teacher Solver | +Noise Schedule | +FAD ↓ | +FD ↓ | +KLD ↓ | +
---|---|---|---|---|---|---|
Unguided | +1 | +DDIM | +Uniform | +13.48 | +45.75 | +2.409 | +
External CFG | +3 | +DDIM | +Uniform | +8.565 | +38.67 | +2.015 | +
Heun | +Karras | +7.421 | +39.36 | +1.976 | +||
CFG Distillation with Fixed Weight |
+ 3 | +Heun | +Karras | +5.702 | +33.18 | +1.494 | +
Uniform | +3.859 | +27.79 | +1.421 | +|||
CFG Distillation with Random Weight |
+ 4 | +Heun | +Uniform | +3.180 | +27.92 | +1.394 | +
6 | +2.975 | +28.63 | +1.378 | +
@article{bai2023accelerating, +@inproceedings{bai2024accelerating, author = {Bai, Yatong and Dang, Trung and Tran, Dung and Koishida, Kazuhito and Sojoudi, Somayeh}, - title = {Accelerating Diffusion-Based Text-to-Audio Generation with Consistency Distillation}, - journal={arXiv preprint arXiv:2309.10740}, - year = {2023} + title = {ConsistencyTTA: Accelerating Diffusion-Based Text-to-Audio Generation with Consistency Distillation}, + booktitle = {INTERSPEECH}, + year = {2024} }