Increasing the 'retry_job_output_collection' value in universe_wsgi.ini fixed the "failed" job problem. Thanks for the suggestion Nate. --Hemant -----Original Message----- From: Nate Coraor [mailto:nate@bx.psu.edu] Sent: Friday, May 11, 2012 3:54 PM To: Kelkar, Hemant Cc: galaxy-dev@lists.bx.psu.edu Subject: Re: [galaxy-dev] Problem related to a job that "failed" On May 11, 2012, at 3:39 PM, Kelkar, Hemant wrote:
All,
This question pertains to a local galaxy install where the jobs are being submitted to a cluster running LSF.
I periodically get an error from a galaxy job ("Job output not returned from cluster") even though the job completes properly on the cluster. In researching this issue our systems administrator discovered that the error seems to be happening because of NFS caching. The problem is that the job finishes on the job node, but the Galaxy server doesn't see that reflected in the output file because of delays in the cache update in NFS. The only solution he discovered was turning the caching off on the file system. In our case it is not possible because it will lead to a performance hit that would be not acceptable on the shared cluster. Most other file systems on the cluster are "NFS" so moving the database/pbs folder to another file system is not an option.
Hi Hemant, You wouldn't need to change attribute caching on the entire cluster, just on the filesystem in which Galaxy is stored on the server on which the Galaxy application runs. If that's not possible either, see the 'retry_job_output_collection' option in universe_wsgi.ini.
I generally work around the NFS cache issue but now it is leading to a second problem. Since the job appears to be in some failed state to galaxy (it shows up as red in history) I can't seem to use the output file (even though it is there and I can see it using the "eye" icon) to move to the next step. The file attribute is set right.
I assume a possible solution may be to reset the "failed" flag on the history item. Would this need to be done in the database? Downloading and then re-uploading the result file (a 25+ GB SAM file in this case) may be a workaround but it is not very practical.
Yes, this would be something to update in the database. The process is something like: 1a. Collect the encoded history_dataset_association ID from the errored dataset's display URL. 1b. Use galaxy-dist/scripts/helper.py -d <encoded_hda_id> to decode the id. 1c. In the database, 'select dataset_id from history_dataset_association where id=<decoded_hda_id>' OR 1. Collect the output dataset id from the command line that was executed and logged in the Galaxy server log 2. In the database, 'update dataset set state='ok' where id=<dataset_id>' --nate
Any ideas/suggestions?
Thanks,
Hemant ___________________________________________________________ Please keep all replies on the list by using "reply all" in your mail client. To manage your subscriptions to this and other Galaxy lists, please use the interface at: