does anyone have Galaxy running with PBS Pro (not torque)?
If so, i'm interested to hear how that went. Cheers Mike. -- Michael Pheasant Software Engineer Queensland Facility for Advanced Bioinformatics Level 5, QBP University of Queensland, QLD 4072 T: +61 (0)7 3346 2070 F: +61 (0)7 3346 2101 www.qfab.org
Michael Pheasant wrote:
If so, i'm interested to hear how that went.
Hi Mike, I'm currently working on generic DRMAA code that should be committed soon. Once this is ready, you should be able to use PBS Pro through FedStage's DRMAA library. --nate
Cheers
Mike.
-- Michael Pheasant Software Engineer Queensland Facility for Advanced Bioinformatics Level 5, QBP University of Queensland, QLD 4072 T: +61 (0)7 3346 2070 F: +61 (0)7 3346 2101 www.qfab.org _______________________________________________ galaxy-dev mailing list galaxy-dev@lists.bx.psu.edu http://lists.bx.psu.edu/listinfo/galaxy-dev
Hi Nate, Nate Coraor wrote, On 07/28/2010 12:53 PM:
I'm currently working on generic DRMAA code that should be committed soon. Once this is ready, you should be able to use PBS Pro through FedStage's DRMAA library.
May I ask 'how soon' ? I'm working on a patch to allow SGE jobs to run under different users with DRMAA_Python. If you're going to change the runners framework, I'd better wait until you have the new DRMAA code ready. Thanks, -gordon
Assaf Gordon wrote:
May I ask 'how soon' ? I'm working on a patch to allow SGE jobs to run under different users with DRMAA_Python. If you're going to change the runners framework, I'd better wait until you have the new DRMAA code ready.
It looks like it's committable now, so I'm writing the documentation for it first. Should be in this afternoon. The code is very similar to the SGE code, so you may not have to change anything. --nate
Thanks, -gordon _______________________________________________ galaxy-dev mailing list galaxy-dev@lists.bx.psu.edu http://lists.bx.psu.edu/listinfo/galaxy-dev
Hi Gordon, Does your plan involve having DRMAA handle the multiuser capability itself? I was under the impression that DRMAA cannot do it: http://gridengine.sunsource.net/issues/show_bug.cgi?id=2507 I've likewise started coding this myself. We run LSF, but I'm hoping there won't be anything cluster-specific about it -- so far I'm just doing (e)uid/(e)gid switching plus plain DRMAA. It's not done yet, though -- jobs are submitted as the appropriate user but galaxy is losing track of them. It would of course be even better if DRMAA did this itself, but I expect there will also be many changes to galaxy required, especially regarding dataset management where the actual files are owned by different cluster users. Nate, do you think this will just be a few relatively minor tweaks to the galaxy code or a big project? Thanks, John On Wed, 28 Jul 2010, Nate Coraor wrote:
Assaf Gordon wrote:
May I ask 'how soon' ? I'm working on a patch to allow SGE jobs to run under different users with DRMAA_Python. If you're going to change the runners framework, I'd better wait until you have the new DRMAA code ready.
It looks like it's committable now, so I'm writing the documentation for it first. Should be in this afternoon.
The code is very similar to the SGE code, so you may not have to change anything.
--nate
Thanks, -gordon _______________________________________________ galaxy-dev mailing list galaxy-dev@lists.bx.psu.edu http://lists.bx.psu.edu/listinfo/galaxy-dev
_______________________________________________ galaxy-dev mailing list galaxy-dev@lists.bx.psu.edu http://lists.bx.psu.edu/listinfo/galaxy-dev
John Brunelle wrote:
Hi Gordon,
Does your plan involve having DRMAA handle the multiuser capability itself? I was under the impression that DRMAA cannot do it:
http://gridengine.sunsource.net/issues/show_bug.cgi?id=2507
I've likewise started coding this myself. We run LSF, but I'm hoping there won't be anything cluster-specific about it -- so far I'm just doing (e)uid/(e)gid switching plus plain DRMAA. It's not done yet, though -- jobs are submitted as the appropriate user but galaxy is losing track of them.
It would of course be even better if DRMAA did this itself, but I expect there will also be many changes to galaxy required, especially regarding dataset management where the actual files are owned by different cluster users.
Nate, do you think this will just be a few relatively minor tweaks to the galaxy code or a big project?
Hi John, I look at it as a fairly big project - for the reasons you've found (permissions) I think Galaxy has to have elevated privileges, and a safe way to do that would need to be found. --nate
Thanks,
John
On Wed, 28 Jul 2010, Nate Coraor wrote:
Assaf Gordon wrote:
May I ask 'how soon' ? I'm working on a patch to allow SGE jobs to run under different users with DRMAA_Python. If you're going to change the runners framework, I'd better wait until you have the new DRMAA code ready.
It looks like it's committable now, so I'm writing the documentation for it first. Should be in this afternoon.
The code is very similar to the SGE code, so you may not have to change anything.
--nate
Thanks, -gordon _______________________________________________ galaxy-dev mailing list galaxy-dev@lists.bx.psu.edu http://lists.bx.psu.edu/listinfo/galaxy-dev
_______________________________________________ galaxy-dev mailing list galaxy-dev@lists.bx.psu.edu http://lists.bx.psu.edu/listinfo/galaxy-dev
Hi John,
Does your plan involve having DRMAA handle the multiuser capability itself? I was under the impression that DRMAA cannot do it:
Indeed, I couldn't find a way to do it inside DRMAA. My patch uses sudo+setuid .
I've likewise started coding this myself. We run LSF, but I'm hoping there won't be anything cluster-specific about it -- so far I'm just doing (e)uid/(e)gid switching plus plain DRMAA.
My plan is exactly that. Have the regular galaxy SGE-runner do everything upto building the DRMAAs' jobTemplate, then, instead of "drmaa.runJob", write the JobTemplate to file (JSON/Pickle/Whatever), and sudo a tiny python script the will setuid() to the right user and execute the job based on the JobTemplate's data. This tiny script will return the SGE-JobID (or an error), and from there on the SGE-Runner module will go on as before. Hopefully this will incur minimum changes to the galaxy code (basically, just calling a different function at /lib/galaxy/jobs/runners/sge.py:217 ). The external script + 'sudo' is required because I don't want to have the Galaxy python script run as root (and without root I can't change the UID back-and-forth between users).
It's not done yet, though -- jobs are submitted as the appropriate user but galaxy is losing track of them.
My plan is to have the script return the job-id, and so galaxy will just pick it up and use it as if it run it directly.
It would of course be even better if DRMAA did this itself, but I expect there will also be many changes to galaxy required, especially regarding dataset management where the actual files are owned by different cluster users.
for starters I will require all users to be part of the 'galaxy' group, and all files will be 'g+rw' . Regards, -gordon
participants (4)
-
Assaf Gordon
-
John Brunelle
-
Michael Pheasant
-
Nate Coraor