-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathinterpol_loss_job
34 lines (28 loc) · 1.23 KB
/
interpol_loss_job
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
#!/bin/sh
#SBATCH --partition=cms-uhh ## or allgpu / cms / cms-uhh / maxgpu
#SBATCH --time=24:00:00
#SBATCH --nodes=1
#SBATCH --job-name loss_interpol # give job unique name
#SBATCH --output ./interpol_loss_log/%j-std.out # terminal output
#SBATCH --error ./interpol_loss_log/%j-std.err
#S BATCH --mail-type END
#S BATCH --mail-user lennart.kaemmle@desy.de
#SBATCH --constraint=GPU
# #SBATCH --nodelist=max-cmsg004 # you can select specific nodes, if necessary
#SBATCH --constraint="P100|V100" # ask for specific types of GPUs
# (GPU drivers, anaconda, .bashrc, etc) --> likely not necessary if you're installing your own cudatoolkit & cuDNN versions
#source /etc/profile.d/modules.sh
#module load maxwell
#module load cuda
#module load anaconda/3
#module load mpi/mpich-x86_64
# module load mpi/openmpi3-x86_64
. ~/.bashrc
export PATH="/home/kaemmlel/miniconda3/envs/torch/bin:$PATH"
set -x
# export PATH="/home/kaemmlel/miniconda3/envs/torch/bin:$PATH"
export PYTHONPATH="" # openmpi module adds python 2 packages to pythonpath
# go to your folder with your python scripts
cd ~/Projects/amazing_ai
# run
time python -m amazing_ai.scripts.interpol_loss --model out/12145719*/ --interpol @interpol_ss_blur_6000_100.h5