diff --git a/dpgen2/flow/dpgen_loop.py b/dpgen2/flow/dpgen_loop.py index 78a4c15b..0ff95abb 100644 --- a/dpgen2/flow/dpgen_loop.py +++ b/dpgen2/flow/dpgen_loop.py @@ -80,7 +80,7 @@ def make_block_optional_parameter(cl_optional_parameter): def make_next_optional_parameter(optional_parameter): return { "data_mixed_type": optional_parameter["data_mixed_type"], - "finetune_mode": "no", # not to do finetune for `next` loop + "finetune_mode": "no", # not to do finetune for `next` loop } @@ -434,7 +434,8 @@ def _loop( "exploration_scheduler" ], "optional_parameter": make_next_optional_parameter( - steps.inputs.parameters["optional_parameter"]), + steps.inputs.parameters["optional_parameter"] + ), "expl_task_grp": scheduler_step.outputs.parameters["expl_task_grp"], } next_step = Step( diff --git a/dpgen2/op/run_dp_train.py b/dpgen2/op/run_dp_train.py index 4710f82e..9ea3eccb 100644 --- a/dpgen2/op/run_dp_train.py +++ b/dpgen2/op/run_dp_train.py @@ -142,9 +142,7 @@ def _make_train_command_old( str(init_model), train_script_name, ] - elif finetune_mode == "finetune" or ( - do_init_model and init_model_with_finetune - ): + elif finetune_mode == "finetune" or (do_init_model and init_model_with_finetune): command = ( dp_command + [