47 lines
1.1 KiB
Bash
47 lines
1.1 KiB
Bash
set -e
|
|
|
|
cd ../../../../
|
|
|
|
cudaid=$1
|
|
root=$2
|
|
dataset=fs_contest_data
|
|
method=fedavg_ft
|
|
outdir=exp_out/${method}
|
|
|
|
if [ ! -d ${outdir} ];then
|
|
mkdir -p ${outdir}
|
|
fi
|
|
|
|
echo "HPO starts..."
|
|
|
|
lrs=(0.01 0.1 0.05 0.005 0.5)
|
|
local_updates=(1 2 4)
|
|
steps=(5 10 15)
|
|
|
|
|
|
for (( i=0; i<${#lrs[@]}; i++ ))
|
|
do
|
|
for (( j=0; j<${#local_updates[@]}; j++ ))
|
|
do
|
|
for (( s=0; s<${#steps[@]}; s++ ))
|
|
do
|
|
log=${outdir}/gin_lr-${lrs[$i]}_step-${local_updates[$j]}_lstep-${steps[$s]}_on_${dataset}.log
|
|
for k in {1..3}
|
|
do
|
|
python federatedscope/main.py --cfg benchmark/B-FHTL/scripts/Grpah-DC/fedavg_gnn_minibatch_on_multi_task.yaml \
|
|
data.root ${root} \
|
|
device ${cudaid} \
|
|
data.type ${dataset} \
|
|
optimizer.lr ${lrs[$i]} \
|
|
federate.local_update_steps ${local_updates[$j]} \
|
|
trainer.finetune.before_eval True \
|
|
trainer.finetune.steps ${steps[$s]} \
|
|
seed $k >>${log} 2>&1
|
|
done
|
|
python federatedscope/parse_exp_results.py --input ${log}
|
|
done
|
|
done
|
|
done
|
|
|
|
echo "HPO ends."
|