From d8f958b74914d1635730029c4bbb0a93e5aa22d7 Mon Sep 17 00:00:00 2001 From: root Date: Sun, 3 Aug 2025 17:17:11 +0000 Subject: [PATCH 1/2] Save Config parameters to wandb in SFT --- verl/trainer/fsdp_sft_trainer.py | 3 +++ 1 file changed, 3 insertions(+) diff --git a/verl/trainer/fsdp_sft_trainer.py b/verl/trainer/fsdp_sft_trainer.py index e597f712b7e..b23aac6e702 100644 --- a/verl/trainer/fsdp_sft_trainer.py +++ b/verl/trainer/fsdp_sft_trainer.py @@ -684,10 +684,13 @@ def fit(self): # TODO: add a unified tracking if rank == 0: + from omegaconf import OmegaConf + tracking = Tracking( project_name=self.config.trainer.project_name, experiment_name=self.config.trainer.experiment_name, default_backend=self.config.trainer.logger, + config=OmegaConf.to_container(self.config, resolve=True), ) global_step = self.resume_global_step # Start from resumed step From 7ad2ff9c34f5b790da0e339159e3b38160c17827 Mon Sep 17 00:00:00 2001 From: root Date: Sun, 3 Aug 2025 18:15:48 +0000 Subject: [PATCH 2/2] move import to the top of the file --- verl/trainer/fsdp_sft_trainer.py | 4 +--- 1 file changed, 1 insertion(+), 3 deletions(-) diff --git a/verl/trainer/fsdp_sft_trainer.py b/verl/trainer/fsdp_sft_trainer.py index b23aac6e702..ca018fb492c 100644 --- a/verl/trainer/fsdp_sft_trainer.py +++ b/verl/trainer/fsdp_sft_trainer.py @@ -30,7 +30,7 @@ import hydra import torch import torch.distributed -from omegaconf import DictConfig +from omegaconf import DictConfig, OmegaConf from peft import LoraConfig, TaskType, get_peft_model from tensordict import TensorDict from torch import nn, optim @@ -684,8 +684,6 @@ def fit(self): # TODO: add a unified tracking if rank == 0: - from omegaconf import OmegaConf - tracking = Tracking( project_name=self.config.trainer.project_name, experiment_name=self.config.trainer.experiment_name,