Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add new LMMS evaluation task for wild vision benchmark #247

Merged
merged 4 commits into from
Sep 13, 2024
Merged

Conversation

Luodian
Copy link
Contributor

@Luodian Luodian commented Sep 12, 2024

This pull request adds a new LMMS evaluation task named "wildvision_0630" for the wild vision benchmark. The task includes specific configuration settings such as dataset name, test split, output type, and prompt details. This task is added to the existing LMMS evaluation tasks for various subjects.

We also update the evaluation metrics (score, win rate...) for wildvision benchmark to align with their updated evaluation logic.

This commit adds a new LMMS evaluation task for the wild vision benchmark. The task is named "wildvision_0630" and includes specific configuration settings such as dataset name, test split, output type, and prompt details. This task is added to the existing LMMS evaluation tasks for various subjects.
@Luodian Luodian merged commit e77fb31 into main Sep 13, 2024
2 checks passed
@Luodian Luodian deleted the dev/fix_tags branch November 23, 2024 14:35
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant