I'm working on unpacking a zip file into multiple datasets.
I think this is the code path
Upload.py
UploadToolAction
upload_common.py:
get_uploaded_datesets
new_upload
new_history_upload or new_library_upload
Then a job gets spooled
Which calles add_file in
data_source/upload.py
And does the expansion of the zip
I can unpack the zip and create files in the dataset's path there.
But I don't know how to create more dataset associations, and I'm not sure that it makes sense to create datasets on the fly in data_source/upload.py .
Should I pass some information along with data_source/upload.py about how to create dataset object and associate them with library/history associations?
Or maybe I can pass in some kind of a callback that can handle the dataset expansion?
(I'm pretty new to python, but it seems similar to ruby)
I thought about a composite dataset, but that seems like overloading that concept. Really the files I'm thinking about uplaoding are 8 independent BAMs or fastqs or whatever – not a set of files that are related to each other.
Any suggestions?
Brad
--
Brad Langhorst
New England Biolabs
langhorst@neb.com