Skip to content

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Summary statistics for ensembles #208

Closed
wkitlasten opened this issue Jan 2, 2021 · 6 comments
Closed

Summary statistics for ensembles #208

wkitlasten opened this issue Jan 2, 2021 · 6 comments

Comments

@wkitlasten
Copy link
Collaborator

I would like to propose including some simple methods to generate concise quantitative metrics for ensembles, similar to those discussed in the publication below. Any thoughts on those or other metrics that should be included?

https://lib.dr.iastate.edu/cgi/viewcontent.cgi?article=1072&context=ge_at_pubs

On a similar note, I am trying to evaluate how well my IES setup (parameterization, weights, etc) reproduces "synthetic" observations of interest (not in the history matching dataset but "known" from the simulation) from a suite of selected realizations. In my mind, the most logical measure is something along the lines of the number of ensemble standard deviations between the ens mean of the obs of interest and the "known" values. Then, if I have 10 tests for a single site I could hopefully infer that the same setup likely captures real world obs of interest within the same number of stds (assuming similar ensemble metrics mentioned above for the obs in the history matching dataset). Thoughts, or references along those lines?

@jtwhite79
Copy link
Collaborator

I think this would be great - any way(s) to help distill down ensemble results is welcome. I guess you would process by observation group (and it seems like this is mostly about time series , right?). I think @smwesten-usgs might has some good metrics and/or ideas on this front that we could borrow from tsproc?

On that idea of comparing history matching results to synthetic "truth" results, I think the paired simple-complex analysis of Doherty and Christensen is one of the best ways b/c it is focused on exposing biases in the relation between synthetic truth obs and history matching results.

@jtwhite79
Copy link
Collaborator

@wkitlasten - Any movement on this? It would be nice contrib...

@wkitlasten
Copy link
Collaborator Author

wkitlasten commented May 28, 2021 via email

@mnfienen
Copy link
Collaborator

@weskitlasten and @jtwhite79 - sorry I missed this issue before. I made a framework for metrics like this and included some of them. It should be easy to add those that are missing in the framework I laid out there. https://github.com/pypest/pyemu/blob/develop/pyemu/utils/metrics.py If I get a few free moments, I'll try and add the ones that are in the paper you mentioned.

@wkitlasten
Copy link
Collaborator Author

Hey @mnfienen, nice work. I have been using the metrics recently. I hope to add something to it in the near future.

@briochh
Copy link
Collaborator

briochh commented Jan 18, 2023

@wkitlasten, moving this to discussions like #164 it'll probs be in the ongoing improvements basket?

@pypest pypest locked and limited conversation to collaborators Jan 18, 2023
@briochh briochh converted this issue into discussion #396 Jan 18, 2023

This issue was moved to a discussion.

You can continue the conversation there. Go to discussion →

Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants