-
Notifications
You must be signed in to change notification settings - Fork 38
/
run_pretrain_multi.sh
37 lines (33 loc) · 1.02 KB
/
run_pretrain_multi.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
#!/bin/bash
expdir=./tmp
# set up environment variables for Torch DistributedDataParallel
WORLD_SIZE_JOB=\$SLURM_NTASKS
RANK_NODE=\$SLURM_NODEID
PROC_PER_NODE=4
MASTER_ADDR_JOB=\$SLURM_SUBMIT_HOST
MASTER_PORT_JOB="12234"
DDP_BACKEND=c10d
HYDRA_FULL_ERROR=1 python -u ./fairseq/fairseq_cli/hydra_train.py \
--config-dir ./fairseq/examples/hubert/config/contentvec \
--config-name contentvec \
hydra.run.dir=${expdir} \
task.data=metadata \
task.label_dir=label \
task.labels=["km"] \
task.spk2info=spk2info.dict \
task.crop=true \
dataset.train_subset=train \
dataset.valid_subset=valid \
dataset.num_workers=10 \
dataset.max_tokens=1230000 \
checkpoint.keep_best_checkpoints=10 \
criterion.loss_weights=[10,1e-5] \
model.label_rate=50 \
model.encoder_layers_1=3 \
model.logit_temp_ctr=0.1 \
model.ctr_layers=[-6] \
model.extractor_mode="default" \
optimization.update_freq=[1] \
optimization.max_update=100000 \
lr_scheduler.warmup_updates=8000 \
2>&1