Hi,


I have configured a new galaxy-project site with SLURM (version 14). I have one server with a Galaxy-Project instance, one node with SLURM server and two SLURM worker nodes. I have compile SLURM-DRMAA from source codes. When I run “drmaa-run /bin/hostname” it’s work. But, when I try to run the server I got the next error:

Traceback (most recent call last):
  File "/home/galaxy-dist/lib/galaxy/webapps/galaxy/buildapp.py", line 39, in app_factory
    app = UniverseApplication( global_conf = global_conf, **kwargs )
  File "/home/galaxy-dist/lib/galaxy/app.py", line 141, in __init__
    self.job_manager = manager.JobManager( self )
  File "/home/galaxy-dist/lib/galaxy/jobs/manager.py", line 23, in __init__
    self.job_handler = handler.JobHandler( app )
  File "/home/galaxy-dist/lib/galaxy/jobs/handler.py", line 32, in __init__
    self.dispatcher = DefaultJobDispatcher( app )
  File "/home/galaxy-dist/lib/galaxy/jobs/handler.py", line 704, in __init__
    self.job_runners = self.app.job_config.get_job_runner_plugins( self.app.config.server_name )
  File "/home/galaxy-dist/lib/galaxy/jobs/__init__.py", line 621, in get_job_runner_plugins
    rval[id] = runner_class( self.app, runner[ 'workers' ], **runner.get( 'kwds', {} ) )
  File "/home/galaxy-dist/lib/galaxy/jobs/runners/drmaa.py", line 81, in __init__
    self.ds.initialize()
  File "/home/galaxy-dist/eggs/drmaa-0.7.6-py2.6.egg/drmaa/session.py", line 257, in initialize
    py_drmaa_init(contactString)
  File "/home/galaxy-dist/eggs/drmaa-0.7.6-py2.6.egg/drmaa/wrappers.py", line 73, in py_drmaa_init
    return _lib.drmaa_init(contact, error_buffer, sizeof(error_buffer))
  File "/home/galaxy-dist/eggs/drmaa-0.7.6-py2.6.egg/drmaa/errors.py", line 151, in error_check
    raise _ERRORS[code - 1](error_string)
AlreadyActiveSessionException: code 11: DRMAA session already exist.
[root@galaxy-project galaxy-dist]#


This is my “job_conf.xml”:

<job_conf>
    <plugins workers="4">
        <plugin id="local" type="runner" load="galaxy.jobs.runners.local:LocalJobRunner"/>
<plugin id="drmaa" type="runner" load="galaxy.jobs.runners.drmaa:DRMAAJobRunner"/>
        <plugin id="cli" type="runner" load="galaxy.jobs.runners.cli:ShellJobRunner" />
        <plugin id="slurm" type="runner" load="galaxy.jobs.runners.slurm:SlurmJobRunner">
            <param id="drmaa_library_path">/usr/local/lib/libdrmaa.so</param>
        </plugin>
    </plugins>
    <handlers>
        <handler id="main"/>
    </handlers>
    <destinations default="drmaa_slurm">
        <destination id="local" runner="local"/>
        <destination id="multicore_local" runner="local">
          <param id="local_slots">4</param>
          <param id="embed_metadata_in_job">True</param>
          <job_metrics /> 
        </destination>
        <destination id="docker_local" runner="local">
          <param id="docker_enabled">true</param>
        </destination>
        <destination id="drmaa_slurm" runner="drmaa">
            <param id="galaxy_external_runjob_script">scripts/drmaa_external_runner.py</param>
            <param id="galaxy_external_killjob_script">scripts/drmaa_external_killer.py</param>
            <param id="galaxy_external_chown_script">scripts/external_chown_script.py</param>
        </destination>
        <destination id="direct_slurm" runner="slurm">
            <param id="nativeSpecification">--time=00:01:00</param>
        </destination>
    </destinations>
    <resources default="default">
      <group id="default"></group>
      <group id="memoryonly">memory</group>
      <group id="all">processors,memory,time,project</group>
    </resources>
    <tools>
        <tool id="foo" handler="trackster_handler">
            <param id="source">trackster</param>
        </tool>
        <tool id="bar" destination="dynamic"/>
        <tool id="longbar" destination="dynamic" resources="all" />
        <tool id="baz" handler="special_handlers" destination="bigmem"/>
    </tools>
    <limits>
        <limit type="registered_user_concurrent_jobs">2</limit>
        <limit type="anonymous_user_concurrent_jobs">1</limit>
        <limit type="destination_user_concurrent_jobs" id="local">1</limit>
        <limit type="destination_user_concurrent_jobs" tag="mycluster">2</limit>
        <limit type="destination_user_concurrent_jobs" tag="longjobs">1</limit>
        <limit type="destination_total_concurrent_jobs" id="local">16</limit>
        <limit type="destination_total_concurrent_jobs" tag="longjobs">100</limit>
        <limit type="walltime">24:00:00</limit>
        <limit type="output_size">10GB</limit>
    </limits>
</job_conf>


Can you help me? I am newbie with “Galaxy Project” administration.




THANKS IN ADVANCE!!!!





Alfonso Pardo Diaz
System Administrator / Researcher
c/ Sola nº 1; 10200 Trujillo, ESPAÑA
Tel: +34 927 65 93 17 Fax: +34 927 32 32 37

CETA-Ciemat logo



----------------------------
Confidencialidad: 
Este mensaje y sus ficheros adjuntos se dirige exclusivamente a su destinatario y puede contener información privilegiada o 
confidencial. Si no es vd. el destinatario indicado, queda notificado de que la utilización, divulgación y/o copia sin autorización está 
prohibida en virtud de la legislación vigente. Si ha recibido este mensaje por error, le rogamos que nos lo comunique 
inmediatamente respondiendo al mensaje y proceda a su destrucción.
 
Disclaimer: 
This message and its attached files is intended exclusively for its recipients and may contain confidential information. If you received 
this e-mail in error you are hereby notified that any dissemination, copy or disclosure of this communication is strictly prohibited and 
may be unlawful. In this case, please notify us by a reply and delete this email and its contents immediately.
----------------------------