Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Clarification on Evaluation Metrics in Your Video-ChatGPT Paper #130

Open
sanchitsingh001 opened this issue Feb 2, 2025 · 0 comments
Open

Comments

@sanchitsingh001
Copy link

Hey,
I am a student researcher at the James Siliibard Brown Center for AI at SDSU. I am really interested in the paper that you wrote, "Video-ChatGPT: Towards Detailed Video Understanding via Large Vision and Language Models.". I got a few questions regarding the evaluation metrics that you use. 

First, you give a score between 1 and 5 using a GPT pipeline. Is it like you input the predicted response and the actual response in the gpt model and have a system prompt saying that give a score to the predicted response based on Correctness of Information, Detail Orientation, Contextual Understanding, Temporal Understanding, and Consistency. Also, if you do this, how do you do this for the whole dataset? I don't see that you mentioned exactly how the evaluation works. 

Also, there is a percentage score associated with the dataset I am referring to in Table 2. I don't understand how you calculate that accuracy for each dataset. Like, what's the exact structure or calculations you use?

Overall I like your paper and approach; I will appreciate it if you can answer the above questions.
Thank you,
Sanchit Singh
Student Researcher at James Sillibard Brown Center for AI at SDSU

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant