Hello All,
This is more of a general policy question. The head of our Research Informatics Support Systems group here at the University of Minnesota wants to know how other institutions are handling the large amount of data generated by next generation sequencing facilities and processed by local Galaxy instances. Specifically, I think she is interested in what sort of data retention policies institutions have in place - for instance how long do you store data, what sort of quotas do you put in place, what sort of backups do you have in place, etc....
Any comments would be appreciated.
Thanks for your time, -John
On Mon, Oct 24, 2011 at 4:53 AM, John Chilton chilton@msi.umn.edu wrote:
Hello All,
This is more of a general policy question. The head of our Research Informatics Support Systems group here at the University of Minnesota wants to know how other institutions are handling the large amount of data generated by next generation sequencing facilities and processed by local Galaxy instances. Specifically, I think she is interested in what sort of data retention policies institutions have in place - for instance how long do you store data, what sort of quotas do you put in place, what sort of backups do you have in place, etc....
Any comments would be appreciated.
Thanks for your time, -John
Might be worth asking on the bioinformatics core facilities mailing list too, I know some of the folk there use Galaxy in their institutes: http://open-bio.org/mailman/listinfo/bioinfo-core
Peter
On Oct 24, 2011, at 2:57 AM, Peter Cock wrote:
On Mon, Oct 24, 2011 at 4:53 AM, John Chilton chilton@msi.umn.edu wrote:
Hello All,
This is more of a general policy question. The head of our Research
Informatics Support Systems group here at the University of Minnesota wants to know how other institutions are handling the large amount of data generated by next generation sequencing facilities and processed by local Galaxy instances. Specifically, I think she is interested in what sort of data retention policies institutions have in place - for instance how long do you store data, what sort of quotas do you put in place, what sort of backups do you have in place, etc....
Any comments would be appreciated.
Thanks for your time, -John
Might be worth asking on the bioinformatics core facilities mailing list too, I know some of the folk there use Galaxy in their institutes: http://open-bio.org/mailman/listinfo/bioinfo-core
Peter
We're in the (very long, drawn-out) process of integrating much of our local bioinformatics resources to also deal with this problem. We have a general plan in place, including adding automated tape backup and a number of other features on the analysis end (storage for fast I/O, etc), but I'm interested in hearing more from others on this.
chris
galaxy-dev@lists.galaxyproject.org