-
Notifications
You must be signed in to change notification settings - Fork 65
/
Copy pathrun448.sh
43 lines (41 loc) · 1.29 KB
/
run448.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
export MASTER_PORT=$((12000 + $RANDOM % 20000))
export OMP_NUM_THREADS=1
echo "PYTHONPATH: ${PYTHONPATH}"
which_python=$(which python)
echo "which python: ${which_python}"
export PYTHONPATH=${PYTHONPATH}:${which_python}
export PYTHONPATH=${PYTHONPATH}:.
echo "PYTHONPATH: ${PYTHONPATH}"
JOB_NAME='videomamba_small_res224to448'
OUTPUT_DIR="$(dirname $0)"
LOG_DIR="./logs/${JOB_NAME}"
PARTITION='video5'
NNODE=1
NUM_GPUS=8
NUM_CPU=128
srun --mpi=pmi2 \
-p ${PARTITION} \
-n${NNODE} \
--gres=gpu:${NUM_GPUS} \
--ntasks-per-node=1 \
--cpus-per-task=${NUM_CPU} \
python -m torch.distributed.launch --nproc_per_node=${NUM_GPUS} --use_env main.py \
--root_dir_train your_imagenet_path/train/ \
--meta_file_train your_imagenet_path/meta/train.txt \
--root_dir_val your_imagenet_path/val/ \
--meta_file_val your_imagenet_path/meta/val.txt \
--model videomamba_small \
--finetune your_model_path/videomamba_small_res224.pth \
--input-size 448 \
--batch-size 64 \
--num_workers 16 \
--lr 5e-6 \
--min-lr 5e-6 \
--weight-decay 1e-8 \
--warmup-epochs 5 \
--epochs 30 \
--drop-path 0.15 \
--no-model-ema \
--output_dir ${OUTPUT_DIR}/ckpt \
--bf16 \
--dist-eval