I accidentally shut down my galaxy server in the middle of an job execution, after that I try to restart my galaxy server, it always shows the error that a job is in new state, and I can't start my galaxy server at all. I guess I have to clear that job out of the queue, so my galaxy server can start. Does anyone know how to clear the job queue? or any idea how to fix the problem? the error log as below: --------------------------------------------------------- galaxy.jobs.runners.local INFO 2009-07-13 17:30:21,924 starting workers galaxy.jobs.runners.local DEBUG 2009-07-13 17:30:21,945 5 workers ready galaxy.jobs.runners.local INFO 2009-07-13 17:30:21,950 starting workers galaxy.jobs.runners.local DEBUG 2009-07-13 17:30:21,969 5 workers ready galaxy.jobs.schedulingpolicy.roundrobin INFO 2009-07-13 17:30:21,992 RoundRobin policy: initialized galaxy.jobs INFO 2009-07-13 17:30:21,992 job scheduler policy is galaxy.jobs.schedulingpolicy.roundrobin:UserRoundRobin galaxy.jobs INFO 2009-07-13 17:30:22,005 job manager started galaxy.jobs DEBUG 2009-07-13 17:30:22,202 no runner: 326 is still in new state, adding to the jobs queue galaxy.jobs DEBUG 2009-07-13 17:30:22,202 no runner: 330 is still in new state, adding to the jobs queue Traceback (most recent call last): File "/media/Applications/work/spring2009/galaxy/galaxy3/lib/galaxy/web/buildapp.py", line 56, in app_factory app = UniverseApplication( global_conf = global_conf, **kwargs ) File "/media/Applications/work/spring2009/galaxy/galaxy3/lib/galaxy/app.py", line 66, in __init__ self.job_manager = jobs.JobManager( self ) File "/media/Applications/work/spring2009/galaxy/galaxy3/lib/galaxy/jobs/__init__.py", line 33, in __init__ self.job_queue = JobQueue( app, self.dispatcher ) File "/media/Applications/work/spring2009/galaxy/galaxy3/lib/galaxy/jobs/__init__.py", line 110, in __init__ self.__check_jobs_at_startup() File "/media/Applications/work/spring2009/galaxy/galaxy3/lib/galaxy/jobs/__init__.py", line 127, in __check_jobs_at_startup job_wrapper = JobWrapper( job, self.app.toolbox.tools_by_id[ job.tool_id ], self ) KeyError: u'xslt_converter_16'
participants (1)
-
R