Question: Best way running OMA using SLURM
0
gravatar for dimar.gonzalez.soto
8 months ago by
dimar.gonzalez.soto0 wrote:

Hi everyone, I'm executing OMA through a cluster using a bash script mentioned in the documentation of OMA standalone. For parallelization I'm using SLURM with this script:

sbatch --array=1-100 -N1 <<EOF
#!/bin/sh
export NR_PROCESSES=100
/usr/local/bin/OMA/bin/OMA
EOF

SLURM is running only one OMA proccess per node, I want to execute all 100 proccess at same time. How can I achieve that?

slurm oma • 297 views
ADD COMMENTlink modified 8 months ago by adrian.altenhoff620 • written 8 months ago by dimar.gonzalez.soto0

Is this all there is to the script? What is your actual OMA command line with options/inputs/outputs?

ADD REPLYlink written 8 months ago by genomax73k

I have not put input/output to the OMA executable

ADD REPLYlink modified 8 months ago • written 8 months ago by dimar.gonzalez.soto0

How will you actually use OMA for analysis of your data if you don't tell it what to analyze.

You seem to have copied the section about SLURM parallelization from OMA Standalone help page but are not actually constructed the OMA command line for running the analysis.

ADD REPLYlink modified 8 months ago • written 8 months ago by genomax73k

There is a default parameter file that is used in case there are no arguments

ADD REPLYlink written 8 months ago by adrian.altenhoff620

I see. @adrian: Thanks for the clarification.

ADD REPLYlink written 8 months ago by genomax73k

Ok I will construct a command properly

ADD REPLYlink written 8 months ago by dimar.gonzalez.soto0
1
gravatar for adrian.altenhoff
8 months ago by
Switzerland
adrian.altenhoff620 wrote:

Hi

your command should indeed start 100 jobs of oma standalone in parallel. however the cluster configuration specifies how much resources a job should be allocated by default if not otherwise specified. I assume that your configuration allows to use a lot of resources and hence it will only start one job per node. This should not be a problem as long as you have enough nodes in your cluster... However, you will block others from using the cluster if you don't specify tight resource limits (see man sbatch; in particular --mem option)

Best wishes Adrian

ADD COMMENTlink written 8 months ago by adrian.altenhoff620

if start one process per node, should be it about slurm configuration?

ADD REPLYlink written 8 months ago by dimar.gonzalez.soto0
Please log in to add an answer.

Help
Access

Use of this site constitutes acceptance of our User Agreement and Privacy Policy.
Powered by Biostar version 2.3.0
Traffic: 1034 users visited in the last hour