... | ... | @@ -24,11 +24,11 @@ For Shared Installations: |
|
|
|
|
|
|
|
|
```
|
|
|
1. mkdir -p </shared/dir/>
|
|
|
1. mkdir -p </shared/dir/bin>
|
|
|
|
|
|
2. mv /var/lib/fahclient </shared/dir/>
|
|
|
2. mv /var/lib/fahclient </shared/dir/bin>
|
|
|
|
|
|
3. mv /usr/bin/FAH* </shared/dir/>
|
|
|
3. mv /usr/bin/FAH* </shared/dir/bin>
|
|
|
```
|
|
|
|
|
|
|
... | ... | @@ -317,31 +317,22 @@ rm -rf ./$SLURM_JOBID |
|
|
[folding@ocf ~/fah]# cat runGPU.sh
|
|
|
|
|
|
#!/bin/bash
|
|
|
|
|
|
# OCF - www.ocf.co.uk
|
|
|
|
|
|
# set the number of nodes
|
|
|
|
|
|
#SBATCH --nodes=1
|
|
|
|
|
|
#SBATCH --ntasks-per-node=1
|
|
|
#SBATCH --gres=gpu:2
|
|
|
#SBATCH --job-name=Folding@Home
|
|
|
#SBATCH --partition=<partition_name>
|
|
|
#SBATCH --time=21-0
|
|
|
|
|
|
#SBATCH --gres=gpu:1
|
|
|
|
|
|
|
|
|
# set name of job
|
|
|
FAHDIR=<fah_directory>
|
|
|
cd $FAHDIR
|
|
|
mkdir $FAHDIR/bin/slurm_work/$SLURM_JOBID
|
|
|
|
|
|
#SBATCH --job-name=folding-at-home-GPU
|
|
|
|
|
|
#SBATCH --partition=spot-folding-gpu
|
|
|
|
|
|
module load cuda/10.1
|
|
|
$FAHDIR/bin/FAHClient --config $FAHDIR/bin/config-gpu.xml --checkpoint 10 --chdir $FAHDIR/bin/slurm_work/$SLURM_JOBID --http-addresses 0:$(shuf -i 10000-11000 -n1) --command-port $(shuf -i 11000-12000 -n1) --log $FAHDIR/output/$SLURM_JOBID.log --smp --gpu --exit-when-done
|
|
|
|
|
|
cd ~/fah
|
|
|
|
|
|
mkdir ./$SLURM_JOBID
|
|
|
|
|
|
./FAHClient --config ./config-gpu.xml --checkpoint 10 --chdir ./$SLURM_JOBID --http-addresses 0:$(shuf -i 10000-11000 -n1) --command-port $(shuf -i 11000-12000 -n1) --log ~/fah/output/$SLURM_JOBID.log --smp --gpu --exit-when-done
|
|
|
|
|
|
rm -rf ./$SLURM_JOBID
|
|
|
rm -rf $FAHDIR/bin/slurm_work/$SLURM_JOBID
|
|
|
``` |
|
|
\ No newline at end of file |