I have the problem that there are a lot of files that are not deleted in the
Among these files I have failed uploads and duplicated library files. The latter do not
correspond to new versions of existing datasets.
I have run the cleanup_datasets.py script but I still have files dating back a year that
are in this condition.
I have no clue as why/how this could have happened.
The galaxy instance is installed on a cluster with the filesystem on a NAS.
I have been running 16.01 on this instance (will update as soon as I can stop the
Could anyone suggest a way to identify those files that have to be in database/files so
that I can delete all the rest??
Thanks in advance!
Show replies by date