-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
43 update re iding process records to handle data objects wo data file description or url or with ID not found in the Database #46
Conversation
…-data-objects-wo-data-file-description-or-url
…-legacy-projects' into 43-update-re-iding-process-records-to-handle-data-objects-wo-data-file-description-or-url
- workflow data with missing DO ids -> separate json output - add option to process-records to update links (or not, for testing) - improve logging
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Added methods to deal with unliked DO's and computing compliant file path. Extended Analysis activity set. Generated new data json dumps.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Implemented method to get has_input for legacy id, add method to fail omics processing record if DO is orphaned, added check for readbased analysis activity type.
This PR provides several updates to the commands in
re_id_tool.py
:failed_record_dump
outfileupdate-links
to process records in order to test locally without filesystem changesAdditionally this PR includes the results and logs for
extract-records
andprocess-records
on:Workflows with an "orphan" data object ID (cannot be found in our DB or in any of the
data_objects.json
are written to a failed record dump, along with the rest of that Workflow's has_output data objects. If the failing Workflow is ReadQC, all workflows are failed.