-
Notifications
You must be signed in to change notification settings - Fork 24
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
full-length datasets #17
Comments
It would be useful to have even a single full length recording, ideally one that includes one of the existing datasets, so the performance difference can be estimated. |
Echoing some comments by @sofroniewn from the gitter chat... Most of the datasets are ~7 minutes (3000 frames @ 6-7 Hz), frame rates are in the But here's the current lineup, and what we could add:
I think my vote would be to standardize all of them @ 8 Hz by downsampling if neccessary, and post everything we have. Though this will increase the data sizes by quite a bit. |
I see. I guess I was thinking of series 01 and 03 which are in fact 1-2 minutes long. Sounds like a good idea to standardize @8hz and post everything. The longest dataset then will be ~10 minutes long. What is the bandwidth limitation, do you have to pay to host the data? Maybe consider also adding a single good recording, with many neurons, for ~1 hour. |
Ok, updates are done! Data set durations are now as follows:
So all are now close to 7-8 Hz, all are at least 7 min, and the longest is 17 min. We've now posted everything we have from the original providers. Will add this info to the website. My only concern adding a ~1 hr dataset to the test data is that its size @ 8 Hz could become onerous for some people's machines / some algorithms, and submitting already requires running algorithms across 7 moderately sized datasets. We could downsample a longer one even more, say to 8000 frames @ 2.5 Hz, but then it wouldn't be consistent with the sample rate of the others. That said, always happy to add extra datasets as training data of any size just for people to play with, storage / bandwidth isn't really an issue. |
Cool, thanks for expanding the datasets! I will be curious if it improves the scores or not. For someone who comes to the website to see which algorithms might be useful, a single full-length dataset would be invaluable for assessing not just accuracy on a typical recording, but also how fast the algorithms are on realistic recordings. Shouldn't any algorithm be able to run on a 2hr dataset @30hz ? Most of our data is in that range. |
The frame rate for the harvey lab datasets is not the same for .00 and .01. The frame rates saved in the .json file for each submitted dataset should be correct (for the .01 it is 3hz). I think having (moderate) diversity in these datasets is a feature, not a bug. The test results so far to me look like a lot of low-rank structure: much better performance on some datasets than others. It's useful to see this to understand the successes and breakdowns of the algorithms under different conditions (and truth definitions). |
Well, I think that the wide range of results on different datasets has solely to do with the different types of ground truth definitions, and very little to do with the algorithms. Which isn't great, this benchmark is supposed to test the algorithms, not the annotation method! |
Current datasets are way too short and not at all representative of real use scenarios. Many more cells will be detected from a typical 1-2 hour recording than from the length of time provided here. These don't have to be downloadable, perhaps only available for running algorithms remotely on?
The text was updated successfully, but these errors were encountered: