Re: [galaxy-user] Export to file
Hi Dave, Yes, if your Galaxy instance is on the internet, for entire history transfer, you can skip the curl download and just enter the URL from the public Main Galaxy server into your Galaxy directly. To load large data over 2G that is local (datasets, not history archives), you can use the data library option. The idea is to load into a library, then move datasets from libraries into histories as needed. Help is in our wiki here: http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Libraries http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Uploading%20Library%20Files Take care, Jen Galaxy team On 10/17/12 3:21 PM, Dave Corney wrote:
Hi Jen,
Thanks for your response and suggestion. Just so that it is clear, for your second method, where I export to file and then use curl, I will download to my computer as an intermediate stage? Is there a simple way to take the history and datasets from PSU galaxy to our Princeton galaxy directly (without downloading to my computer first)? Unfortunately, we don't have FTP on our own galaxy, which is why I was looking for alternatives (each file is >2GB, so uploading through the browser won't work either). It seems that to import from file, the file needs to have a URL and I'm not sure how to go about that if the file is store locally on my computer.
Thanks, Dave
On Wed, Oct 17, 2012 at 6:12 PM, Jennifer Jackson <jen@bx.psu.edu <mailto:jen@bx.psu.edu>> wrote:
Hi Dave,
To export larger files, you can use a different method. Open up a terminal window on your computer and type in at the prompt ($):
$ curl -0 '<file_link>' > name_the_output
Where <file_link> can be obtained by right-clicking on the disc icon for the dataset and selecting "Copy link location".
If you are going to import into a local Galaxy, exporting entire histories, or a history comprised of datasets that you have copied/grouped together, may be a quick alternative. From the history panel, use "Options (gear icon) -> Export to File" to generate a link, then use curl again to perform the download. The "Import from File" function (in the same menu) can be used in your local Galaxy to incorporate the history and the datasets it contains.
Hopefully this helps, but please let us know if you have more questions,
Jen Galaxy team
On 10/17/12 2:37 PM, Dave Corney wrote:
Hi list,
Is there a currently a known problem with the "export to file" function? I'm trying to migrate some data from the public galaxy to a private one; the export function worked well with a small (~100mb) dataset, but it has not been working with larger datasets (>2GB) and I get the error: Server Error. An error occurred. See the error logs for more information. (Turn debug on to display exception reports here). Is there a limit on the file size of the export? If so, what is it?
Thanks in advance, Dave
_____________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org <http://usegalaxy.org>. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/__listinfo/galaxy-dev <http://lists.bx.psu.edu/listinfo/galaxy-dev>
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
-- Jennifer Jackson http://galaxyproject.org
-- Jennifer Jackson http://galaxyproject.org
Dave, There's likely something problematic about your history that causing problems. Can you share with me the history that's generating the error? To do so, from the history options menu --> Share/Publish --> Share with a User --> my email address Thanks, J. On Oct 17, 2012, at 6:58 PM, Jennifer Jackson wrote:
Hi Dave,
Yes, if your Galaxy instance is on the internet, for entire history transfer, you can skip the curl download and just enter the URL from the public Main Galaxy server into your Galaxy directly.
To load large data over 2G that is local (datasets, not history archives), you can use the data library option. The idea is to load into a library, then move datasets from libraries into histories as needed. Help is in our wiki here: http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Libraries http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Uploading%20Library%20Files
Take care,
Jen Galaxy team
On 10/17/12 3:21 PM, Dave Corney wrote:
Hi Jen,
Thanks for your response and suggestion. Just so that it is clear, for your second method, where I export to file and then use curl, I will download to my computer as an intermediate stage? Is there a simple way to take the history and datasets from PSU galaxy to our Princeton galaxy directly (without downloading to my computer first)? Unfortunately, we don't have FTP on our own galaxy, which is why I was looking for alternatives (each file is >2GB, so uploading through the browser won't work either). It seems that to import from file, the file needs to have a URL and I'm not sure how to go about that if the file is store locally on my computer.
Thanks, Dave
On Wed, Oct 17, 2012 at 6:12 PM, Jennifer Jackson <jen@bx.psu.edu <mailto:jen@bx.psu.edu>> wrote:
Hi Dave,
To export larger files, you can use a different method. Open up a terminal window on your computer and type in at the prompt ($):
$ curl -0 '<file_link>' > name_the_output
Where <file_link> can be obtained by right-clicking on the disc icon for the dataset and selecting "Copy link location".
If you are going to import into a local Galaxy, exporting entire histories, or a history comprised of datasets that you have copied/grouped together, may be a quick alternative. From the history panel, use "Options (gear icon) -> Export to File" to generate a link, then use curl again to perform the download. The "Import from File" function (in the same menu) can be used in your local Galaxy to incorporate the history and the datasets it contains.
Hopefully this helps, but please let us know if you have more questions,
Jen Galaxy team
On 10/17/12 2:37 PM, Dave Corney wrote:
Hi list,
Is there a currently a known problem with the "export to file" function? I'm trying to migrate some data from the public galaxy to a private one; the export function worked well with a small (~100mb) dataset, but it has not been working with larger datasets (>2GB) and I get the error: Server Error. An error occurred. See the error logs for more information. (Turn debug on to display exception reports here). Is there a limit on the file size of the export? If so, what is it?
Thanks in advance, Dave
_____________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org <http://usegalaxy.org>. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/__listinfo/galaxy-dev <http://lists.bx.psu.edu/listinfo/galaxy-dev>
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
-- Jennifer Jackson http://galaxyproject.org
-- Jennifer Jackson http://galaxyproject.org ___________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/listinfo/galaxy-dev
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
Hi Jeremy, Thanks for your offer of help. By the time I got your email I had already added many new jobs to the history that are either running now or waiting to run. Since I read somewhere that if the history is running then there are problems exporting I shared a clone of the history with you. The clone should be identical to the history that I was having problems with yesterday. I can share with you the original history once the jobs have finished running (but it might take a while). Thanks, Dave On Wed, Oct 17, 2012 at 10:35 PM, Jeremy Goecks <jeremy.goecks@emory.edu>wrote:
Dave,
There's likely something problematic about your history that causing problems. Can you share with me the history that's generating the error? To do so, from the history options menu --> Share/Publish --> Share with a User --> my email address
Thanks, J.
On Oct 17, 2012, at 6:58 PM, Jennifer Jackson wrote:
Hi Dave,
Yes, if your Galaxy instance is on the internet, for entire history transfer, you can skip the curl download and just enter the URL from the public Main Galaxy server into your Galaxy directly.
To load large data over 2G that is local (datasets, not history archives), you can use the data library option. The idea is to load into a library, then move datasets from libraries into histories as needed. Help is in our wiki here: http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Libraries
http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Uploading%20Library%20Files
Take care,
Jen Galaxy team
On 10/17/12 3:21 PM, Dave Corney wrote:
Hi Jen,
Thanks for your response and suggestion. Just so that it is clear, for your second method, where I export to file and then use curl, I will download to my computer as an intermediate stage? Is there a simple way to take the history and datasets from PSU galaxy to our Princeton galaxy directly (without downloading to my computer first)? Unfortunately, we don't have FTP on our own galaxy, which is why I was looking for alternatives (each file is >2GB, so uploading through the browser won't work either). It seems that to import from file, the file needs to have a URL and I'm not sure how to go about that if the file is store locally on my computer.
Thanks, Dave
On Wed, Oct 17, 2012 at 6:12 PM, Jennifer Jackson <jen@bx.psu.edu <mailto:jen@bx.psu.edu>> wrote:
Hi Dave,
To export larger files, you can use a different method. Open up a terminal window on your computer and type in at the prompt ($):
$ curl -0 '<file_link>' > name_the_output
Where <file_link> can be obtained by right-clicking on the disc icon for the dataset and selecting "Copy link location".
If you are going to import into a local Galaxy, exporting entire histories, or a history comprised of datasets that you have copied/grouped together, may be a quick alternative. From the history panel, use "Options (gear icon) -> Export to File" to generate a link, then use curl again to perform the download. The "Import from File" function (in the same menu) can be used in your local Galaxy to incorporate the history and the datasets it contains.
Hopefully this helps, but please let us know if you have more
questions,
Jen Galaxy team
On 10/17/12 2:37 PM, Dave Corney wrote:
Hi list,
Is there a currently a known problem with the "export to file" function? I'm trying to migrate some data from the public galaxy to a private one; the export function worked well with a small (~100mb) dataset, but it has not been working with larger datasets (>2GB) and I get the error: Server Error. An error occurred. See the error logs for more information. (Turn debug on to display exception reports here). Is there a limit on the file size of the export? If so, what is it?
Thanks in advance, Dave
_____________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org <http://usegalaxy.org>. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/__listinfo/galaxy-dev <http://lists.bx.psu.edu/listinfo/galaxy-dev>
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
-- Jennifer Jackson http://galaxyproject.org
-- Jennifer Jackson http://galaxyproject.org ___________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/listinfo/galaxy-dev
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
I've reworked the code to handle large history export files in -central changeset afc8e9345268., and this should solve your issue. This change should make it out to our public server this coming week. Best, J. On Oct 18, 2012, at 12:36 PM, Dave Corney wrote:
Hi Jeremy,
Thanks for your offer of help. By the time I got your email I had already added many new jobs to the history that are either running now or waiting to run. Since I read somewhere that if the history is running then there are problems exporting I shared a clone of the history with you. The clone should be identical to the history that I was having problems with yesterday. I can share with you the original history once the jobs have finished running (but it might take a while).
Thanks, Dave
On Wed, Oct 17, 2012 at 10:35 PM, Jeremy Goecks <jeremy.goecks@emory.edu> wrote: Dave,
There's likely something problematic about your history that causing problems. Can you share with me the history that's generating the error? To do so, from the history options menu --> Share/Publish --> Share with a User --> my email address
Thanks, J.
On Oct 17, 2012, at 6:58 PM, Jennifer Jackson wrote:
Hi Dave,
Yes, if your Galaxy instance is on the internet, for entire history transfer, you can skip the curl download and just enter the URL from the public Main Galaxy server into your Galaxy directly.
To load large data over 2G that is local (datasets, not history archives), you can use the data library option. The idea is to load into a library, then move datasets from libraries into histories as needed. Help is in our wiki here: http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Libraries http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Uploading%20Library%20Files
Take care,
Jen Galaxy team
On 10/17/12 3:21 PM, Dave Corney wrote:
Hi Jen,
Thanks for your response and suggestion. Just so that it is clear, for your second method, where I export to file and then use curl, I will download to my computer as an intermediate stage? Is there a simple way to take the history and datasets from PSU galaxy to our Princeton galaxy directly (without downloading to my computer first)? Unfortunately, we don't have FTP on our own galaxy, which is why I was looking for alternatives (each file is >2GB, so uploading through the browser won't work either). It seems that to import from file, the file needs to have a URL and I'm not sure how to go about that if the file is store locally on my computer.
Thanks, Dave
On Wed, Oct 17, 2012 at 6:12 PM, Jennifer Jackson <jen@bx.psu.edu <mailto:jen@bx.psu.edu>> wrote:
Hi Dave,
To export larger files, you can use a different method. Open up a terminal window on your computer and type in at the prompt ($):
$ curl -0 '<file_link>' > name_the_output
Where <file_link> can be obtained by right-clicking on the disc icon for the dataset and selecting "Copy link location".
If you are going to import into a local Galaxy, exporting entire histories, or a history comprised of datasets that you have copied/grouped together, may be a quick alternative. From the history panel, use "Options (gear icon) -> Export to File" to generate a link, then use curl again to perform the download. The "Import from File" function (in the same menu) can be used in your local Galaxy to incorporate the history and the datasets it contains.
Hopefully this helps, but please let us know if you have more questions,
Jen Galaxy team
On 10/17/12 2:37 PM, Dave Corney wrote:
Hi list,
Is there a currently a known problem with the "export to file" function? I'm trying to migrate some data from the public galaxy to a private one; the export function worked well with a small (~100mb) dataset, but it has not been working with larger datasets (>2GB) and I get the error: Server Error. An error occurred. See the error logs for more information. (Turn debug on to display exception reports here). Is there a limit on the file size of the export? If so, what is it?
Thanks in advance, Dave
_____________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org <http://usegalaxy.org>. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/__listinfo/galaxy-dev <http://lists.bx.psu.edu/listinfo/galaxy-dev>
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
-- Jennifer Jackson http://galaxyproject.org
-- Jennifer Jackson http://galaxyproject.org ___________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/listinfo/galaxy-dev
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
Hi Jeremy, That's really wonderful - thanks so much for taking the time and effort to do this! When you say large history, is there a size limit that I should be aware of, or will it handle anything that my quota can accept? Thanks, Dave On Sat, Oct 20, 2012 at 2:44 PM, Jeremy Goecks <jeremy.goecks@emory.edu>wrote:
I've reworked the code to handle large history export files in -central changeset afc8e9345268., and this should solve your issue. This change should make it out to our public server this coming week.
Best, J.
On Oct 18, 2012, at 12:36 PM, Dave Corney wrote:
Hi Jeremy,
Thanks for your offer of help. By the time I got your email I had already added many new jobs to the history that are either running now or waiting to run. Since I read somewhere that if the history is running then there are problems exporting I shared a clone of the history with you. The clone should be identical to the history that I was having problems with yesterday. I can share with you the original history once the jobs have finished running (but it might take a while).
Thanks, Dave
On Wed, Oct 17, 2012 at 10:35 PM, Jeremy Goecks <jeremy.goecks@emory.edu>wrote:
Dave,
There's likely something problematic about your history that causing problems. Can you share with me the history that's generating the error? To do so, from the history options menu --> Share/Publish --> Share with a User --> my email address
Thanks, J.
On Oct 17, 2012, at 6:58 PM, Jennifer Jackson wrote:
Hi Dave,
Yes, if your Galaxy instance is on the internet, for entire history transfer, you can skip the curl download and just enter the URL from the public Main Galaxy server into your Galaxy directly.
To load large data over 2G that is local (datasets, not history archives), you can use the data library option. The idea is to load into a library, then move datasets from libraries into histories as needed. Help is in our wiki here: http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Libraries
http://wiki.g2.bx.psu.edu/Admin/Data%20Libraries/Uploading%20Library%20Files
Take care,
Jen Galaxy team
On 10/17/12 3:21 PM, Dave Corney wrote:
Hi Jen,
Thanks for your response and suggestion. Just so that it is clear, for your second method, where I export to file and then use curl, I will download to my computer as an intermediate stage? Is there a simple way to take the history and datasets from PSU galaxy to our Princeton
galaxy
directly (without downloading to my computer first)? Unfortunately, we don't have FTP on our own galaxy, which is why I was looking for alternatives (each file is >2GB, so uploading through the browser won't work either). It seems that to import from file, the file needs to have a URL and I'm not sure how to go about that if the file is store locally on my computer.
Thanks, Dave
On Wed, Oct 17, 2012 at 6:12 PM, Jennifer Jackson <jen@bx.psu.edu <mailto:jen@bx.psu.edu>> wrote:
Hi Dave,
To export larger files, you can use a different method. Open up a terminal window on your computer and type in at the prompt ($):
$ curl -0 '<file_link>' > name_the_output
Where <file_link> can be obtained by right-clicking on the disc icon for the dataset and selecting "Copy link location".
If you are going to import into a local Galaxy, exporting entire histories, or a history comprised of datasets that you have copied/grouped together, may be a quick alternative. From the history panel, use "Options (gear icon) -> Export to File" to generate a link, then use curl again to perform the download. The "Import from File" function (in the same menu) can be used in your local Galaxy to incorporate the history and the datasets it contains.
Hopefully this helps, but please let us know if you have more questions,
Jen Galaxy team
On 10/17/12 2:37 PM, Dave Corney wrote:
Hi list,
Is there a currently a known problem with the "export to file" function? I'm trying to migrate some data from the public galaxy to a private one; the export function worked well with a small (~100mb) dataset, but it has not been working with larger datasets (>2GB) and I get the error: Server Error. An error occurred. See the error logs for more information. (Turn debug on to display exception reports here). Is there a limit on the file size of the export? If so, what is it?
Thanks in advance, Dave
_____________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org <http://usegalaxy.org>. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/__listinfo/galaxy-dev <http://lists.bx.psu.edu/listinfo/galaxy-dev>
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
-- Jennifer Jackson http://galaxyproject.org
-- Jennifer Jackson http://galaxyproject.org ___________________________________________________________ The Galaxy User list should be used for the discussion of Galaxy analysis and other features on the public server at usegalaxy.org. Please keep all replies on the list by using "reply all" in your mail client. For discussion of local Galaxy instances and the Galaxy source code, please use the Galaxy Development list:
http://lists.bx.psu.edu/listinfo/galaxy-dev
To manage your subscriptions to this and other Galaxy lists, please use the interface at:
participants (3)
-
Dave Corney
-
Jennifer Jackson
-
Jeremy Goecks