use_gpu: True device: 0 outdir: graph_exp/pfedhpo seed: 12345 early_stop: patience: 100000 improve_indicator_mode: mean federate: mode: 'standalone' make_global_eval: False total_round_num: 4000 sample_client_rate: 1.0 share_local_model: False use_diff: True data: root: data/ type: mini-graph-dc dataloader: type: pyg model: task: graph type: gin hidden: 64 personalization: local_param: ['encoder_atom', 'encoder', 'clf'] train: batch_or_epoch: epoch local_update_steps: 1 optimizer: type: SGD trainer: type: graphminibatch_trainer eval: freq: 1 metrics: ['acc', 'correct'] count_flops: False split: ['train', 'val', 'test'] hpo: scheduler: pfedhpo num_workers: 0 pfedhpo: use: True discrete: True train_fl: False train_anchor: False target_fl_total_round: 400 ss: 'scripts/example_configs/pfedhpo/mini_graph_dc/ss.yaml' metric: 'client_summarized_weighted_avg.val_avg_loss' working_folder: graph_exp/pfedhpo/working