Solved!!!!! The problem was I have configured the job_conf.xml two plugins entry with DRMAA. I have deleted the entry: <plugin id="slurm" type="runner" load="galaxy.jobs.runners.slurm:SlurmJobRunner"> And now works! Thanks El 25/09/2014, a las 08:12, Pardo Diaz, Alfonso <alfonso.pardo@ciemat.es<mailto:alfonso.pardo@ciemat.es>> escribió: Hi, I have configured a new galaxy-project site with SLURM (version 14). I have one server with a Galaxy-Project instance, one node with SLURM server and two SLURM worker nodes. I have compile SLURM-DRMAA from source codes. When I run “drmaa-run /bin/hostname” it’s work. But, when I try to run the server I got the next error: Traceback (most recent call last): File "/home/galaxy-dist/lib/galaxy/webapps/galaxy/buildapp.py", line 39, in app_factory app = UniverseApplication( global_conf = global_conf, **kwargs ) File "/home/galaxy-dist/lib/galaxy/app.py", line 141, in __init__ self.job_manager = manager.JobManager( self ) File "/home/galaxy-dist/lib/galaxy/jobs/manager.py", line 23, in __init__ self.job_handler = handler.JobHandler( app ) File "/home/galaxy-dist/lib/galaxy/jobs/handler.py", line 32, in __init__ self.dispatcher = DefaultJobDispatcher( app ) File "/home/galaxy-dist/lib/galaxy/jobs/handler.py", line 704, in __init__ self.job_runners = self.app.job_config.get_job_runner_plugins( self.app.config.server_name ) File "/home/galaxy-dist/lib/galaxy/jobs/__init__.py", line 621, in get_job_runner_plugins rval[id] = runner_class( self.app, runner[ 'workers' ], **runner.get( 'kwds', {} ) ) File "/home/galaxy-dist/lib/galaxy/jobs/runners/drmaa.py", line 81, in __init__ self.ds.initialize() File "/home/galaxy-dist/eggs/drmaa-0.7.6-py2.6.egg/drmaa/session.py", line 257, in initialize py_drmaa_init(contactString) File "/home/galaxy-dist/eggs/drmaa-0.7.6-py2.6.egg/drmaa/wrappers.py", line 73, in py_drmaa_init return _lib.drmaa_init(contact, error_buffer, sizeof(error_buffer)) File "/home/galaxy-dist/eggs/drmaa-0.7.6-py2.6.egg/drmaa/errors.py", line 151, in error_check raise _ERRORS[code - 1](error_string) AlreadyActiveSessionException: code 11: DRMAA session already exist. [root@galaxy-project galaxy-dist]# This is my “job_conf.xml”: <job_conf> <plugins workers="4"> <plugin id="local" type="runner" load="galaxy.jobs.runners.local:LocalJobRunner"/> <plugin id="drmaa" type="runner" load="galaxy.jobs.runners.drmaa:DRMAAJobRunner"/> <plugin id="cli" type="runner" load="galaxy.jobs.runners.cli:ShellJobRunner" /> <plugin id="slurm" type="runner" load="galaxy.jobs.runners.slurm:SlurmJobRunner"> <param id="drmaa_library_path">/usr/local/lib/libdrmaa.so</param> </plugin> </plugins> <handlers> <handler id="main"/> </handlers> <destinations default="drmaa_slurm"> <destination id="local" runner="local"/> <destination id="multicore_local" runner="local"> <param id="local_slots">4</param> <param id="embed_metadata_in_job">True</param> <job_metrics /> </destination> <destination id="docker_local" runner="local"> <param id="docker_enabled">true</param> </destination> <destination id="drmaa_slurm" runner="drmaa"> <param id="galaxy_external_runjob_script">scripts/drmaa_external_runner.py</param> <param id="galaxy_external_killjob_script">scripts/drmaa_external_killer.py</param> <param id="galaxy_external_chown_script">scripts/external_chown_script.py</param> </destination> <destination id="direct_slurm" runner="slurm"> <param id="nativeSpecification">--time=00:01:00</param> </destination> </destinations> <resources default="default"> <group id="default"></group> <group id="memoryonly">memory</group> <group id="all">processors,memory,time,project</group> </resources> <tools> <tool id="foo" handler="trackster_handler"> <param id="source">trackster</param> </tool> <tool id="bar" destination="dynamic"/> <tool id="longbar" destination="dynamic" resources="all" /> <tool id="baz" handler="special_handlers" destination="bigmem"/> </tools> <limits> <limit type="registered_user_concurrent_jobs">2</limit> <limit type="anonymous_user_concurrent_jobs">1</limit> <limit type="destination_user_concurrent_jobs" id="local">1</limit> <limit type="destination_user_concurrent_jobs" tag="mycluster">2</limit> <limit type="destination_user_concurrent_jobs" tag="longjobs">1</limit> <limit type="destination_total_concurrent_jobs" id="local">16</limit> <limit type="destination_total_concurrent_jobs" tag="longjobs">100</limit> <limit type="walltime">24:00:00</limit> <limit type="output_size">10GB</limit> </limits> </job_conf> Can you help me? I am newbie with “Galaxy Project” administration. THANKS IN ADVANCE!!!! Alfonso Pardo Diaz System Administrator / Researcher c/ Sola nº 1; 10200 Trujillo, ESPAÑA Tel: +34 927 65 93 17 Fax: +34 927 32 32 37 [CETA-Ciemat logo]<http://www.ceta-ciemat.es/> ---------------------------- Confidencialidad: Este mensaje y sus ficheros adjuntos se dirige exclusivamente a su destinatario y puede contener información privilegiada o confidencial. Si no es vd. el destinatario indicado, queda notificado de que la utilización, divulgación y/o copia sin autorización está prohibida en virtud de la legislación vigente. Si ha recibido este mensaje por error, le rogamos que nos lo comunique inmediatamente respondiendo al mensaje y proceda a su destrucción. Disclaimer: This message and its attached files is intended exclusively for its recipients and may contain confidential information. If you received this e-mail in error you are hereby notified that any dissemination, copy or disclosure of this communication is strictly prohibited and may be unlawful. In this case, please notify us by a reply and delete this email and its contents immediately. ---------------------------- ___________________________________________________________ Please keep all replies on the list by using "reply all" in your mail client. To manage your subscriptions to this and other Galaxy lists, please use the interface at: http://lists.bx.psu.edu/ To search Galaxy mailing lists use the unified search at: http://galaxyproject.org/search/mailinglists/