I'm working on unpacking a zip file into multiple datasets.
I think this is the code path Upload.py UploadToolAction
upload_common.py: get_uploaded_datesets new_upload new_history_upload or new_library_upload
Then a job gets spooled Which calles add_file in data_source/upload.py
And does the expansion of the zip
I can unpack the zip and create files in the dataset's path there.
But I don't know how to create more dataset associations, and I'm not sure that it makes sense to create datasets on the fly in data_source/upload.py .
Should I pass some information along with data_source/upload.py about how to create dataset object and associate them with library/history associations? Or maybe I can pass in some kind of a callback that can handle the dataset expansion? (I'm pretty new to python, but it seems similar to ruby)
I thought about a composite dataset, but that seems like overloading that concept. Really the files I'm thinking about uplaoding are 8 independent BAMs or fastqs or whatever – not a set of files that are related to each other.
Any suggestions?
Brad -- Brad Langhorst New England Biolabs langhorst@neb.com