Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add initial CITATION.cff file #164

Merged
merged 1 commit into from
Jan 20, 2024

Conversation

ml-evs
Copy link
Contributor

@ml-evs ml-evs commented Jul 27, 2022

re: #156.

Just added @ardunn as the author for this repo now. Zenodo has a separate "contributors" field but not sure how this maps to the CFF schema (probably just as additional authors - not sure how you would feel about that!).

The main goal from #156 is to automatically archive releases to Zenodo. This file will automatically be used by Zenodo to populate the metadata, otherwise for each release, Zenodo will scrape the repo metadata and include all contributors as authors, and will also miss the related DOIs of the paper, and any additional info you want to provide (ORCID, custom titles etc.).

Outstanding issue: there's no satisfactory way to update the CITATION.cff with the latest versioned Zenodo DOI. Can either just use the "concept DOI" for all versions in the citation file, then describe how to cite the versioned DOI in the text, or try to set up some additional procedure that would use the Zenodo API to add one commit after release that bumps the versioned DOI.

repository-code: https://github.com/materialsproject/matbench
url: https://matbench.materialsproject.org/
license: MIT
preferred-citation:
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This "preferred-citation" field is used by GitHub to populate the "Cite this repository" button in the repo.

@ml-evs
Copy link
Contributor Author

ml-evs commented Sep 29, 2022

@ardunn are you still interested in setting up these automated Zenodo archives? Having this file would be a start (the metadata for the first release would be edited versus this file, then we can update it with all benchmark authors going forward). Unfortunately Zenodo archives can currently only have 1 editor (for now: zenodo/zenodo#810).

@ardunn
Copy link
Collaborator

ardunn commented Sep 29, 2022

@ardunn are you still interested in setting up these automated Zenodo archives? Having this file would be a start (the metadata for the first release would be edited versus this file, then we can update it with all benchmark authors going forward). Unfortunately Zenodo archives can currently only have 1 editor (for now: zenodo/zenodo#810).

Hey @ml-evs ! Thanks for the work on this!

Could you explain to me what the advantages of this are compared to just telling people to cite the original paper?

Would it be like, when we release updated or new benchmarks (which we are looking into right now for generative learning and adaptive design) the citations for each would be better organized by version?

I'm just trying to briefly weigh the maintenance effort of adding another account vs the reward of having the citations more neatly accounted for. I'm not against merging this in at all, just wanted a bit more to clarify the advantages

@ml-evs
Copy link
Contributor Author

ml-evs commented Sep 29, 2022

@ardunn are you still interested in setting up these automated Zenodo archives? Having this file would be a start (the metadata for the first release would be edited versus this file, then we can update it with all benchmark authors going forward). Unfortunately Zenodo archives can currently only have 1 editor (for now: zenodo/zenodo#810).

Hey @ml-evs ! Thanks for the work on this!

Could you explain to me what the advantages of this are compared to just telling people to cite the original paper?

Couple of advantages off the top of my head, as discussed in #156. People can cite versions of the leaderboard, so when they make some comparison between algorithms you can immediately see what was on the leaderboard at that time (and how out-of-date the citation is). I've seen a few papers that use matbench by just comparing to the initial published "leaderboard" (before we even discussed setting up this web version!) It should also incentivize people to add benchmarks pre-publication, as they can then cite the benchmark that includes their own data.

Would it be like, when we release updated or new benchmarks (which we are looking into right now for generative learning and adaptive design) the citations for each would be better organized by version?

Yep, as the benchmarks themselves evolve this would also capture that.

I'm just trying to briefly weigh the maintenance effort of adding another account vs the reward of having the citations more neatly accounted for. I'm not against merging this in at all, just wanted a bit more to clarify the advantages

Understood! Once it is set up, it should be pretty painless. Could easily forgo the part of adjusting the citation file as new models are submitted, at which point it just becomes an automated archival of this repo to Zenodo every time you hit release. Happy to chip in with maintenance and future dev if there are other volunteers but understand if this project isn't a priority anymore!

@ardunn
Copy link
Collaborator

ardunn commented Sep 30, 2022

@ml-evs Actually, I am tentatively handing this project off to @hrushikesh-s , a new graduate student in our group. He's looking to work on some new methods of benchmarking alternative paradigms of materials discovery such as adaptive design and generative models, but he's also getting up to speed with maintaining the repo... so this PR might be a useful way for him to dip his toes in. @hrushikesh-s , do you have any thoughts on this?

@ardunn ardunn merged commit 5d7abc5 into materialsproject:main Jan 20, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants