-
Notifications
You must be signed in to change notification settings - Fork 0
/
slurm_tf_tmpl.txt
23 lines (21 loc) · 1.02 KB
/
slurm_tf_tmpl.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
#!/bin/bash
#SBATCH -J ${sample.zfill(2)}samples # Name for your job
#SBATCH --array=5-35
#SBATCH -n 16 # Number of tasks when using MPI. Default is 1, and SLURM assumes the usage of 1 cpu per task.
#SBATCH -N 1 # Number of nodes to spread cores across - default is 1 - if you are not using MPI this should likely be 1
#SBATCH --mem ${mem} # Megabytes of memory requested
#SBATCH -t ${time} # Runtime in minutes
#SBATCH -p ${queue} # Partition to submit to the standard compute node partition(defq) or the express node partition(express)
#SBATCH -o %A_${sample.zfill(2)}samples_%2a_snr.out.txt # Standard output
#SBATCH -e %A_${sample.zfill(2)}samples_%2a_snr.err.txt # Standard error
date
echo "Hello from $(hostname)."
echo "SNR = ${"${"}SLURM_ARRAY_TASK_ID}"
lscpu
module load Anaconda3
source activate $HOME/condatf2
export HDF5_USE_FILE_LOCKING='FALSE'
echo "Run a Python script"
python3 ${py_file_0} ${"${"}SLURM_ARRAY_TASK_ID} ${sample}
python3 ${py_file_1} ${"${"}SLURM_ARRAY_TASK_ID} ${sample}
echo "Ended batch processing at `date`."