Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

JP-3789 Add more FITS comparison support for regression tests #9082

Open
wants to merge 61 commits into
base: main
Choose a base branch
from

Conversation

penaguerrero
Copy link
Contributor

Resolves JP-3789

Closes #

This PR addresses the initial changes we would like to implement in FitsDiff.

Tasks

  • request a review from someone specific, to avoid making the maintainers review every PR
  • add a build milestone, i.e. Build 11.3 (use the latest build if not sure)
  • Does this PR change user-facing code / API? (if not, label with no-changelog-entry-needed)
    • write news fragment(s) in changes/: echo "changed something" > changes/<PR#>.<changetype>.rst (see below for change types)
    • update or add relevant tests
    • update relevant docstrings and / or docs/ page
    • start a regression test and include a link to the running job (click here for instructions)
      • Do truth files need to be updated ("okified")?
        • after the reviewer has approved these changes, run okify_regtests to update the truth files
  • if a JIRA ticket exists, make sure it is resolved properly
news fragment change types...
  • changes/<PR#>.general.rst: infrastructure or miscellaneous change
  • changes/<PR#>.docs.rst
  • changes/<PR#>.stpipe.rst
  • changes/<PR#>.datamodels.rst
  • changes/<PR#>.scripts.rst
  • changes/<PR#>.fits_generator.rst
  • changes/<PR#>.set_telescope_pointing.rst
  • changes/<PR#>.pipeline.rst

stage 1

  • changes/<PR#>.group_scale.rst
  • changes/<PR#>.dq_init.rst
  • changes/<PR#>.emicorr.rst
  • changes/<PR#>.saturation.rst
  • changes/<PR#>.ipc.rst
  • changes/<PR#>.firstframe.rst
  • changes/<PR#>.lastframe.rst
  • changes/<PR#>.reset.rst
  • changes/<PR#>.superbias.rst
  • changes/<PR#>.refpix.rst
  • changes/<PR#>.linearity.rst
  • changes/<PR#>.rscd.rst
  • changes/<PR#>.persistence.rst
  • changes/<PR#>.dark_current.rst
  • changes/<PR#>.charge_migration.rst
  • changes/<PR#>.jump.rst
  • changes/<PR#>.clean_flicker_noise.rst
  • changes/<PR#>.ramp_fitting.rst
  • changes/<PR#>.gain_scale.rst

stage 2

  • changes/<PR#>.assign_wcs.rst
  • changes/<PR#>.badpix_selfcal.rst
  • changes/<PR#>.msaflagopen.rst
  • changes/<PR#>.nsclean.rst
  • changes/<PR#>.imprint.rst
  • changes/<PR#>.background.rst
  • changes/<PR#>.extract_2d.rst
  • changes/<PR#>.master_background.rst
  • changes/<PR#>.wavecorr.rst
  • changes/<PR#>.srctype.rst
  • changes/<PR#>.straylight.rst
  • changes/<PR#>.wfss_contam.rst
  • changes/<PR#>.flatfield.rst
  • changes/<PR#>.fringe.rst
  • changes/<PR#>.pathloss.rst
  • changes/<PR#>.barshadow.rst
  • changes/<PR#>.photom.rst
  • changes/<PR#>.pixel_replace.rst
  • changes/<PR#>.resample_spec.rst
  • changes/<PR#>.residual_fringe.rst
  • changes/<PR#>.cube_build.rst
  • changes/<PR#>.extract_1d.rst
  • changes/<PR#>.resample.rst

stage 3

  • changes/<PR#>.assign_mtwcs.rst
  • changes/<PR#>.mrs_imatch.rst
  • changes/<PR#>.tweakreg.rst
  • changes/<PR#>.skymatch.rst
  • changes/<PR#>.exp_to_source.rst
  • changes/<PR#>.outlier_detection.rst
  • changes/<PR#>.tso_photometry.rst
  • changes/<PR#>.stack_refs.rst
  • changes/<PR#>.align_refs.rst
  • changes/<PR#>.klip.rst
  • changes/<PR#>.spectral_leak.rst
  • changes/<PR#>.source_catalog.rst
  • changes/<PR#>.combine_1d.rst
  • changes/<PR#>.ami.rst

other

  • changes/<PR#>.wfs_combine.rst
  • changes/<PR#>.white_light.rst
  • changes/<PR#>.cube_skymatch.rst
  • changes/<PR#>.engdb_tools.rst
  • changes/<PR#>.guider_cds.rst

@penaguerrero penaguerrero requested a review from a team as a code owner January 17, 2025 19:32
@penaguerrero penaguerrero marked this pull request as draft January 17, 2025 19:32
Copy link

codecov bot commented Jan 17, 2025

Codecov Report

Attention: Patch coverage is 5.91806% with 620 lines in your changes missing coverage. Please review.

Project coverage is 76.69%. Comparing base (6298414) to head (351f32a).
Report is 1 commits behind head on main.

Files with missing lines Patch % Lines
jwst/regtest/st_fitsdiff.py 5.73% 592 Missing ⚠️
jwst/scripts/stfitsdiff.py 9.67% 28 Missing ⚠️
Additional details and impacted files
@@            Coverage Diff             @@
##             main    #9082      +/-   ##
==========================================
+ Coverage   73.66%   76.69%   +3.02%     
==========================================
  Files         368      509     +141     
  Lines       36393    46399   +10006     
==========================================
+ Hits        26809    35584    +8775     
- Misses       9584    10815    +1231     
Flag Coverage Δ *Carryforward flag
nightly 77.33% <ø> (?) Carriedforward from 79fa2a9

*This pull request uses carry forward flags. Click here to find out more.

☔ View full report in Codecov by Sentry.
📢 Have feedback on the report? Share it here.

# Get the number of NaN in each array
nans = [np.isnan(a).size, np.isnan(b).size]
# Calculate stats
values = np.abs(np.abs(anonan) - np.abs(bnonan))
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Do you want to take the absolute values of a and b before taking the difference? This would make 4 .0 and -4.0 equal

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Ah! good catch! yeah, nope. Definitely an oops.

@melanieclarke
Copy link
Collaborator

Thanks for working on this! A couple requests:

  • It would be helpful to have a command-line script to call the new fitsdiff (st_fitsdiff?), especially for testing the various options.
  • In the original ticket, we'd talked about having allowing a separate tolerance for header and data differences. I don't see that here, unless I missed it? I think that would still be helpful.

Also, I started testing by replacing the regular FITSDiff with STFITSDiff in in regtest jwst/regtest/test_nirspec_mos_spec2.py:

#from astropy.io.fits.diff import FITSDiff
from jwst.regtest.st_fitsdiff import STFITSDiff as FITSDiff

This turned up an error for most of the tests:

        stats = {'mean_value_in_a': np.mean(anonan),
                 'mean_value_in_b': np.mean(bnonan),
>                'max_abs_diff': max(values),
                 'min_abs_diff': min(values),
                 'mean_abs_diff': np.mean(values),
                 'std_dev_abs_diff': np.std(values),
                 'max_rel_diff': max(relative_values),
                 'min_rel_diff': min(values / np.abs(bnonan)),
                 'mean_rel_diff': np.mean(relative_values),
                 'std_dev_rel_diff': np.std(relative_values)}
E       ValueError: max() iterable argument is empty

@penaguerrero
Copy link
Contributor Author

thanks @melanieclarke! I added the script so if you pip install it now the command stfitsdiff should work from the terminal.

@penaguerrero penaguerrero marked this pull request as ready for review February 4, 2025 00:32
@penaguerrero penaguerrero requested a review from a team as a code owner February 4, 2025 00:32
@penaguerrero penaguerrero marked this pull request as draft February 4, 2025 00:33
@penaguerrero
Copy link
Contributor Author

I ran regression tests and it seems the failures are not associated with this PR. https://github.com/spacetelescope/RegressionTests/actions/runs/13247429755

@penaguerrero penaguerrero marked this pull request as ready for review February 10, 2025 20:49
@penaguerrero
Copy link
Contributor Author

@melanieclarke
Copy link
Collaborator

running new regression tests: https://github.com/spacetelescope/RegressionTests/actions/runs/13504092700

A couple weird reports in here. See test_niriss_image_detector1[rate]: jwst.regtest.test_niriss_image:

  Difference of a from b:
  * Absolute number differences are too large.
     0.0_abs ..... 0.0  %
  * Relative number differences are too large.
     0.0_rel ..... 0.0  %

What does this mean?

@penaguerrero
Copy link
Contributor Author

ah! the code prints this when the percentage error is the same value for all thresholds... but I didn't think of the case when it is all 0%, in which case the test should not have failed... I'll investigate!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants