-
Notifications
You must be signed in to change notification settings - Fork 25
Description
We had SignalP working nicely in Galaxy on our old instance running on the cluster as the Galaxy user, but on our new Galaxy instance running on the same cluster as the associated user's Linux account this can happen:
Fatal error: Exit code 1 ()
open: can't stat file
apparent state: unit 3 named input.how
lately reading sequential formatted external IO
One or more tasks failed, e.g. 1 from 'signalp -short -t euk /tmp/86666.1.ln.q/tmpJo4GgL/signalp.0.tmp > /tmp/86666.1.ln.q/tmpJo4GgL/signalp.0.tmp.out' gave:
error running HOW
Error 256 from SignalP:
python /mnt/du-synology/v1shr1/galaxy/shed_tools/toolshed.g2.bx.psu.edu/repos/peterjc/tmhmm_and_signalp/7de64c8b258d/tmhmm_and_signalp/tools/protein_analysis/signalp3.py euk 0 8 /mnt/du-synology/v1shr1/galaxy/galaxy-dist/database/job_working_directory/003/3599/galaxy_dataset_7109.dat.fasta.tmp /mnt/du-synology/v1shr1/galaxy/galaxy-dist/database/job_working_directory/003/3599/galaxy_dataset_7109.dat.tabular.tmp
Error is being raised here:
sys.exit("One or more tasks failed, e.g. %i from %r gave:\n%s" % (error_level, cmd, output), |
My script signalp3.py
breaks up the input FASTA file into chunks of 500 sequences and by default uses four worker threads at once calling SignalP (which is single threaded).
This is on top of the optional Galaxy parallelisation setting which breaks up the parent FASTA input file into chunks of 2000 sequences (i.e. 4 times 500):
<parallelism method="basic" split_inputs="fasta_file" split_mode="to_size" split_size="2000" merge_outputs="tabular_file"></parallelism> |
I've not pinned it down but think it is something about SignalP using predictable temp file names clashing when running child processes on a cluster node (and we expect sets of four jobs to get started around the same time on the same nodes).