Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[help] Rerunning Experiments for Replicating Paper Results #22

Open
kevin3567 opened this issue Dec 7, 2024 · 1 comment
Open

[help] Rerunning Experiments for Replicating Paper Results #22

kevin3567 opened this issue Dec 7, 2024 · 1 comment
Labels
help wanted Extra attention is needed

Comments

@kevin3567
Copy link

For paper replication purposes, in the Readme, it is stated that for "Every metric was collected by running the experiment 10 times separately and calculating the average value." Is this done only for collecting the training/inference speed/gpu-usage, or is this also applicable for getting the reported task-specific (i.e. arc-e, boolq, etc.) accuracy scores?

@mikecovlee
Copy link
Member

We collected these performance metrics by running the evaluation 10 separate times and calculating the average values. For accuracy scores, running multiple trials is unnecessary because our code ensures reproducibility when the random seed is fixed in the same environment and on the same device.

@mikecovlee mikecovlee changed the title Rerunning Experiments for Replicating Paper Results [Reproduction] Rerunning Experiments for Replicating Paper Results Dec 8, 2024
@mikecovlee mikecovlee added the help wanted Extra attention is needed label Dec 8, 2024
@mikecovlee mikecovlee changed the title [Reproduction] Rerunning Experiments for Replicating Paper Results [help] Rerunning Experiments for Replicating Paper Results Dec 21, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

2 participants