Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why can't I reproduce your results, and at the same time, there is a significant difference in FLOPS? #102

Closed
oahcang opened this issue May 28, 2024 · 6 comments

Comments

@oahcang
Copy link

oahcang commented May 28, 2024

No description provided.

@jameslahm
Copy link
Collaborator

Thanks for your interest! Could you please provide more details?

@oahcang
Copy link
Author

oahcang commented May 28, 2024

Thanks for your interest! Could you please provide more details?

Under the same conditions, I tested your v10n model to be 8.6 while in Performance it is 6.7, V10s is 24.8

@jameslahm
Copy link
Collaborator

Does this issue #13 help?

@obisidianouc
Copy link

Thanks for your interest! Could you please provide more details?

Under the same conditions, I tested your v10n model to be 8.6 while in Performance it is 6.7, V10s is 24.8

I find the same issue. The inference time of v10n is about 8, and of v8n is about 6 on A100 (huggingface ZeroGPU)

@jameslahm
Copy link
Collaborator

Thanks for your interest! Could you please provide the details of the benchmark and the results?

I find the same issue. The inference time of v10n is about 8, and of v8n is about 6 on A100 (huggingface ZeroGPU)

@jameslahm
Copy link
Collaborator

Please feel free to reopen this issue if you have further questions.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants