Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any plan for support automatic reasoning models size - like which is actually size of the non-full-scaled model #3

Open
foreverpiano opened this issue Feb 21, 2025 · 4 comments
Labels
question Further information is requested

Comments

@foreverpiano
Copy link

Any plan for support automatic reasoning models size - like which is actually size of the non-full-scaled model

@gaocegege
Copy link
Member

Sounds reasonable. Which model are you using?

@gaocegege gaocegege added the question Further information is requested label Feb 21, 2025
@foreverpiano
Copy link
Author

foreverpiano commented Feb 21, 2025

Like DeepSeek? I heard that some APIs will sometimes use the Qwen distilled version. Can u detect that through benchmark scores?

@gaocegege
Copy link
Member

Do you suggest adding models like DeepSeek: R1 Distill Qwen 32B?

@foreverpiano
Copy link
Author

yes

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants