slurm configuration for cromwell server
1
2
Entering edit mode
2.0 years ago
Eugene A ▴ 180

So far, I can describe my question only in general words, but hopefully someone can help me further narrow it down. I am trying to run modified version of GATK pipeline through cromwell in a server mode on a slurm cluster.

So far I setup a slurm cluster with one node, as a test environment and everything works reasonably well: I'm able to push several instances of my pipeline in parallel to the cromwell server and slurm creates a nice queue of tasks. The problem is, that currently I am hitting the limit of IO operations on the disk where cromwell server is running, which slows everything down:

enter image description here

Here is the result of 18 pipelines submission with 1h interval. As can be seen I/O operations are at it's limit restricting the overall performance.

Now I'd like to add some more nodes to Slum to deal with it. _BUT_ from what I learn so far - all slurm nodes will use the same shared disk, meaning that I will not gain any performance increase from additional computational resources. Is it correct? Or can I force slurm to use local hard drives for each node? If so, how I can configure cromwell server and nodes to correctly copy files back and force?

The question is quite vague but I thank the main idea is clear - ho to increase performance over what I have by addition more nodes to the cluster (if it is possible at all)

Best wishes, Eugene

cromwell slurm wdl • 1.1k views
ADD COMMENT
1
Entering edit mode

are you using a Lustre file system?

ADD REPLY
0
Entering edit mode

Hi, sorry for slow response, I was educating myself a bit on the topic) No, currently we do not use any special fs, but seems that it is unavoidable, thanks! What is your experince with lustre?

ADD REPLY
0
Entering edit mode

It is essentially abouth the same issue described here https://github.com/broadinstitute/cromwell/issues/5802 also witout any replies :(

ADD REPLY
1
Entering edit mode
23 months ago
Eugene A ▴ 180

maybe someone later will find it usefull: looks like it easier to equip server with a nmve disk and run such server as a single-node cluster fully using its reaourses by ajusting amount of parallel instances of piplene.

ADD COMMENT

Login before adding your answer.

Traffic: 1787 users visited in the last hour
Help About
FAQ
Access RSS
API
Stats

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.

Powered by the version 2.3.6