diff --git a/.gitattributes b/.gitattributes index 1ef325f1b111266a6b26e0196871bd78baa8c2f3..6d8d55dcefee422ccb00d231cff75af2b4a97b86 100644 --- a/.gitattributes +++ b/.gitattributes @@ -57,3 +57,12 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text # Video files - compressed *.mp4 filter=lfs diff=lfs merge=lfs -text *.webm filter=lfs diff=lfs merge=lfs -text +checkpoints/archer_Llama3-8B-I_strategic/tokenizer.json filter=lfs diff=lfs merge=lfs -text +checkpoints/archer_Llama3-8B-I_word/tokenizer.json filter=lfs diff=lfs merge=lfs -text +wandb/run-20250914_072202-c40vakf9/run-c40vakf9.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250914_191149-a5wf2wi9/run-a5wf2wi9.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250915_014504-agz3jw75/run-agz3jw75.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250915_111603-tnh8ytpw/run-tnh8ytpw.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250915_170938-mlb6ufl5/run-mlb6ufl5.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250915_234149-solrky9k/run-solrky9k.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250916_001517-sdm2bc8f/run-sdm2bc8f.wandb filter=lfs diff=lfs merge=lfs -text diff --git a/.vscode/launch.json b/.vscode/launch.json new file mode 100644 index 0000000000000000000000000000000000000000..fd1c636b459ac32d88e8ff4c0a934ffc0648dc92 --- /dev/null +++ b/.vscode/launch.json @@ -0,0 +1,61 @@ +{ + "version": "0.2.0", + "configurations": [ + { + "name": "OfflineArcher RSA Training", + "type": "python", + "request": "launch", + "python": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "program": "${workspaceFolder}/main.py", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "env": { + "PYTORCH_CUDA_ALLOC_CONF": "expandable_segments:True", + "ACCELERATE_USE_DEEPSPEED": "true", + "MASTER_PORT": "29500", + "TMPDIR": "$HOME/tmp", + "NCCL_P2P_DISABLE": "1", + "CUDA_LAUNCH_BLOCKING": "1", + "CUDA_DEVICE_MAX_CONNECTIONS": "1", + "PYTHONUNBUFFERED": "1", + "NCCL_DEBUG": "INFO", + "NCCL_DEBUG_SUBSYS": "INIT,ENV" + }, + "deepSpeed": { + "enable": true, + "configPath": "${env:HOME}/codes/ForesightOptim/configs/deepspeed_zero3.yaml" + }, + "console": "integratedTerminal", + "justMyCode": false, + "cwd": "${workspaceFolder}" + } + ] +} \ No newline at end of file diff --git a/Algorithms.py b/Algorithms.py new file mode 100644 index 0000000000000000000000000000000000000000..badd2eab5fa8b22f76365dcd1a4149ffc13f5a2f --- /dev/null +++ b/Algorithms.py @@ -0,0 +1,755 @@ +import lightning as L +from lightning.pytorch.utilities import rank_zero_only +import torch +import os +import gc +torch.set_float32_matmul_precision("high") + +from SimulateOnEnv import batch_simulate_on_environment +from lightning.pytorch.callbacks import Callback +from transformers import AutoModelForCausalLM, AutoTokenizer, AutoConfig +from typing import Optional +from peft import LoraConfig, TaskType, get_peft_model, PeftModel + +from safetensors import safe_open +from safetensors.torch import save_file + +def safe_load(path): + """安全加载权重,处理大小不匹配""" + result = {} + with safe_open(path, framework="pt") as f: + for key in f.keys(): + try: + result[key] = f.get_tensor(key) + except Exception as e: + print(f"Error loading {key}: {str(e)}") + return result + + +def set_special_tokens(model, tokenizer): + if tokenizer.pad_token is None and tokenizer.pad_token_id is None: + print_rank_0(f"[WARNING] the pad token of the tokenizer is None") + # We do not resize the vocab embedding, since it ruins the KL value with the ref_model + tokenizer.pad_token_id = tokenizer.eos_token_id + tokenizer.pad_token = tokenizer.eos_token + # tokenizer.pad_token = tokenizer.decode(0) + + model.config.pad_token_id = tokenizer.pad_token_id + model.config.bos_token_id = tokenizer.bos_token_id + model.config.eos_token_id = tokenizer.eos_token_id + + return model, tokenizer + +def load_model_and_tokenizer(model_name_or_path, actor_checkpoint=None): + + model = AutoModelForCausalLM.from_pretrained( + model_name_or_path, + trust_remote_code=True, + use_cache=False, + torch_dtype=torch.bfloat16, + low_cpu_mem_usage=True, + ) + + if hasattr(model, "ref_model"): + del model.ref_model + + lora_config = LoraConfig( + r=8, + lora_alpha=16, + target_modules=["q_proj", "v_proj"], + lora_dropout=0.05, + bias="none", + task_type=TaskType.CAUSAL_LM, + ) + model = get_peft_model(model, lora_config) + + if actor_checkpoint is not None: + weight_map = {} + with safe_open(actor_checkpoint, framework="pt") as f: + for key in f.keys(): + new_key = key.replace("base_model.model.", "") + weight_map[new_key] = f.get_tensor(key) + + # 应用权重 + for name, param in model.named_parameters(): + for key, tensor in weight_map.items(): + if key in name and param.shape == tensor.shape: + param.data.copy_(tensor) + print(f"加载权重: {name} <- {key}") + break + + tokenizer = AutoTokenizer.from_pretrained( + model_name_or_path, + padding_side="left", # for batch decode + truncation_side="left", + model_max_length=1024, + trust_remote_code=True, + ) + + model.gradient_checkpointing_enable() + model, tokenizer = set_special_tokens(model, tokenizer) + + return model, tokenizer + + +class ActorModel(torch.nn.Module): + def __init__(self, get_device, model_name_or_path, actor_checkpoint=None): + super().__init__() + self.get_device = get_device + self.model, self.tokenizer = load_model_and_tokenizer(model_name_or_path, actor_checkpoint) + + def forward(self, observation, do_sample=True): + obs_ids = self.tokenizer( + observation, + return_tensors="pt", + padding=True, + truncation=True, + max_length=512, + ).to(self.model.device) + obs_embeds = self.model.get_input_embeddings()(obs_ids["input_ids"]) + outputs = self.model.generate( + inputs_embeds=obs_embeds, + attention_mask=obs_ids["attention_mask"], + max_new_tokens=32, + do_sample=do_sample, + pad_token_id=self.tokenizer.eos_token_id, + ) + action = self.tokenizer.batch_decode(outputs, skip_special_tokens=True) + return action + + def behavioral_cloning_loss(self, observation, action, **kwargs): + logsum_probs = self.get_logsum_prob( + observation, action + ) # this line has been refactored and not tested + loss = -logsum_probs.mean() + return loss, {"behavioral_cloning/loss": loss.detach()} + + def get_logsum_prob(self, observation, action_from_dataloader, **kwargs): + action = [a + self.tokenizer.eos_token for a in action_from_dataloader] + alltext = [obs + a for obs, a in zip(observation, action)] + generated_probabilities = self.to_tokens_and_logprobs(alltext) + assert ( + len(generated_probabilities) + == len(alltext) + == len(observation) + == len(action) + ) + mask = torch.zeros_like(generated_probabilities.detach(), dtype=torch.bool) + + for i, (obs, act, text) in enumerate(zip(observation, action, alltext)): + assert text == obs + act + act_ids = self.tokenizer(act, return_tensors="pt", padding=True) + txt_ids = self.tokenizer(text, return_tensors="pt", padding=True) + n_token_act = len( + act_ids["input_ids"][0] + ) # [0] because the batch is one inside the foor loop + n_token_txt = len(txt_ids["input_ids"][0]) + mask[i, n_token_txt - n_token_act - 1 : n_token_txt - 1] = ( + True # the -1 shift is due to the the generated probabilities being shifted + ) + + generated_probabilities = torch.where(mask, generated_probabilities, 1.0) + log_probs = torch.where( + mask, torch.log(generated_probabilities), 0.0 + ) # must be separate from the line above for numerical stability (cannot take log(0.0)) + logsum_probs = torch.sum(log_probs, dim=1) + del act_ids, txt_ids, log_probs, generated_probabilities + return logsum_probs + + def to_tokens_and_logprobs(self, input_texts): + input_ids = self.tokenizer( + input_texts, padding=True, truncation=True, return_tensors="pt" + ).input_ids.to(self.get_device()) + outputs = self.model(input_ids) + probs = torch.softmax(outputs.logits, dim=-1) + + # collect the probability of the generated token -- probability at index 0 corresponds to the token at index 1 + probs = probs[:, :-1, :] + input_ids = input_ids[:, 1:] + gen_probs = torch.gather(probs, 2, input_ids[:, :, None]).squeeze(-1) + del outputs, probs + torch.cuda.empty_cache() + gc.collect() + torch.cuda.memory._set_allocator_settings('max_split_size_mb:32') + return gen_probs + + +class RobertaCritic(torch.nn.Module): + def __init__( + self, + get_device, + discount_factor: float, + tau: float, + expectile: float, + from_checkpoint=None, + ): + super().__init__() + + self.get_device = get_device + self.discount_factor = discount_factor + self.tau = tau + self.expectile = expectile + + ### Define the Critic + from ArcherCritic import ArcherDoubleCritic + + self.critic = ArcherDoubleCritic(in_dim=768, out_dim=1) + self.target_critic = ArcherDoubleCritic(in_dim=768, out_dim=1) + self.soft_update_target_critic(1) + + if from_checkpoint is not None: + checkpoint = torch.load(from_checkpoint, map_location=torch.device("cpu")) + weights = { + k.removeprefix("critic."): v + for k, v in checkpoint["state_dict"].items() + if k.startswith("critic.") + } + self.load_state_dict(weights) + print( + "I have initialized the critic from the checkpoint: ", from_checkpoint + ) + + ### Miscellaneus Shortcuts + self.softmax = torch.nn.Softmax(dim=-1) + self.td_criterion = torch.nn.MSELoss() + self.expectile_criterion = lambda diff: self.loss_value_diff( + diff=diff, expectile=self.expectile + ) + + def get_q(self, observation, action, detach_model=False): + return self.critic.get_q(observation, action, detach_model=detach_model) + + def get_v(self, inputs, detach_model=False): + return self.critic.get_v(inputs, detach_model=detach_model) + + def get_target_v(self, inputs, detach_model=False): + return self.target_critic.get_v(inputs, detach_model=detach_model) + + def get_target_q(self, observation, action, detach_model=False): + return self.target_critic.get_q(observation, action, detach_model=detach_model) + + def get_advantages(self, observation, action): + q1, q2 = self.get_q(observation, action) + v1, v2 = self.get_v(observation) + q = torch.minimum(q1, q2) + v = torch.minimum(v1, v2) + advantages = q - v + return advantages + + def argmax_advantage(self, observation, get_available_actions): + argmax_actions = [] + for obs in observation: + available_actions = get_available_actions(obs) + advantages = torch.as_tensor( + [self.get_advantages([obs], [action]) for action in available_actions] + ) + action = available_actions[torch.argmax(advantages)] + argmax_actions.append(action) + return argmax_actions + + def soft_update_target_critic(self, tau=None): + if tau == None: + tau = self.tau + for target_param, param in zip( + self.target_critic.parameters(), self.critic.parameters() + ): + target_param.data.copy_(target_param.data * (1.0 - tau) + param.data * tau) + + def iql_loss(self, observation, action, reward, next_observation, done, **kwargs): + ### Fitting the Q function + q1, q2 = self.get_q(observation, action, detach_model=False) + q1 = q1.flatten() + q2 = q2.flatten() + + reward = torch.Tensor(reward) # .to(self.agent.device) + done = torch.Tensor(done) # .to(self.agent.device) + + with torch.no_grad(): + target_v1, target_v2 = self.get_target_v(next_observation) + + target_v1 = ( + reward + + torch.logical_not(done) * target_v1.flatten() * self.discount_factor + ) + target_v2 = ( + reward + + torch.logical_not(done) * target_v2.flatten() * self.discount_factor + ) + + q1_loss = self.td_criterion(q1, target_v1) + q2_loss = self.td_criterion(q2, target_v2) + + ### Fitting the value function + with torch.no_grad(): + target_q1, target_q2 = self.get_target_q( + observation, action, detach_model=False + ) + target_q1 = target_q1.flatten() + target_q2 = target_q2.flatten() + + v1, v2 = self.get_v(observation, detach_model=False) + v1 = v1.flatten() + v2 = v2.flatten() + + v1_loss = self.expectile_criterion(diff=target_q1.detach() - v1) + v2_loss = self.expectile_criterion(diff=target_q2.detach() - v2) + + loss = q1_loss + q2_loss + v1_loss + v2_loss + + ### Log and print what's happening + log = self.get_log( + q1=q1, + q2=q2, + v1=v1, + v2=v2, + q1_loss=q1_loss, + q2_loss=q2_loss, + v1_loss=v1_loss, + v2_loss=v2_loss, + target_q1=target_q1, + target_q2=target_q2, + ) + return loss, log + + def loss_value_diff(self, diff, expectile): + """Loss function for iql expectile value difference.""" + weight = torch.where(diff > 0, expectile, (1 - expectile)) + return (weight * (diff**2)).mean() + + def get_log( + self, q1, q2, v1, v2, q1_loss, q2_loss, v1_loss, v2_loss, target_q1, target_q2 + ): + return { + "critic/q1.loss": q1_loss.detach(), + "critic/q2.loss": q2_loss.detach(), + "critic/v1.loss": v1_loss.detach(), + "critic/v2.loss": v2_loss.detach(), + "critic/q1.mean": torch.mean(q1).detach(), + "critic/q1.min": torch.min(q1).detach(), + "critic/q1.max": torch.max(q1).detach(), + "critic/q2.mean": torch.mean(q2).detach(), + "critic/q2.max": torch.max(q2).detach(), + "critic/q2.min": torch.min(q2).detach(), + "critic/v1.mean": torch.mean(v1).detach(), + "critic/v1.min": torch.min(v1).detach(), + "critic/v1.max": torch.max(v1).detach(), + "critic/v2.mean": torch.mean(v2).detach(), + "critic/v2.max": torch.max(v2).detach(), + "critic/v2.min": torch.min(v2).detach(), + "critic/target_q1.mean": torch.mean(target_q1).detach(), + "critic/target_q1.min": torch.min(target_q1).detach(), + "critic/target_q1.max": torch.max(target_q1).detach(), + "critic/target_q2.mean": torch.mean(target_q2).detach(), + "critic/target_q2.max": torch.max(target_q2).detach(), + "critic/target_q2.min": torch.min(target_q2).detach(), + } + + +class Agent(L.LightningModule): + def validation_step(self, batch, batch_idx): + + # Perform evaluation on environment with stochastic policy + return None + eval_dataset = batch_simulate_on_environment( + policy=lambda obs: self.forward(obs), env=None + ) + self.log( + "eval/avg_return", eval_dataset.mean_trajectory_return(), sync_dist=True + ) + self.log( + "eval/std_return", eval_dataset.std_trajectory_return(), sync_dist=True + ) + + # Perform evaluation on environment with deterministic policy + deterministic_eval_dataset = batch_simulate_on_environment( + policy=lambda obs: self.forward(obs, do_sample=False), + env=None, + ) + self.log( + "eval/avg_return_deterministic", + deterministic_eval_dataset.mean_trajectory_return(), + sync_dist=True, + ) + self.log( + "eval/std_return_deterministic", + deterministic_eval_dataset.std_trajectory_return(), + sync_dist=True, + ) + + return eval_dataset.mean_trajectory_return() + + +class BehaviouralCloning(Agent): + def __init__(self, lr: float): + super().__init__() # Initialize LLM base class + self.save_hyperparameters() + + ### Config + self.lr = lr + + ### Initialization + self.agent = ActorModel(get_device=lambda: self.device) + + def forward(self, observation, **kwargs): + return self.agent.forward(observation, **kwargs) + + def training_step(self, batch, batch_idx): + loss, log = self.agent.behavioral_cloning_loss(**batch) + self.log_dict(log, sync_dist=True) + return loss + + def configure_optimizers(self): + from torch.optim import Adam + + # 收集所有需要优化的参数 + optimizer_params = [ + {"params": self.actor.model.parameters(), "lr": self.actor_lr}, + ] + + # 如果需要优化critic,添加其参数 + if self.optimize_critic: + optimizer_params.append({ + "params": self.critic.critic.parameters(), + "lr": self.critic_lr + }) + + # 创建单个优化器 + optimizer = Adam(optimizer_params) + + return optimizer + + +class FilteredBehaviouralCloning(BehaviouralCloning): + def __init__(self, lr: float, filter: float): + super().__init__(lr) + + self.filter = filter + + def configure_callbacks(self): + return FilterDataset(filter=self.filter) + + +class FilterDataset(Callback): + def __init__(self, filter: float): + self.filter = filter + + def on_fit_start(self, trainer, algorithm): + print("*** Filtering Dataset ***") + dataset = trainer.datamodule.dataset + print("Statistics of Input Dataset") + print("Number of Trajectories:", dataset.nTrajectories()) + print("Number of Trajectories:", len(dataset)) + dataset.keep_top_fraction_of_trajectories(fraction=self.filter) + trainer.datamodule.dataset = dataset + print("Statistics of Filtered Dataset") + print("Number of Trajectories:", dataset.nTrajectories()) + print("Number of Trajectories:", len(dataset)) + + +class ActorCritic(Agent): + def __init__( + self, + model_name_or_path: str, + actor_lr: float, + critic_lr: float, + tau: float, + accumulate_grad_batches: int, + discount_factor: float, + critic_expectile: float, + optimize_critic: bool, + actor_checkpoint=None, + critic_checkpoint=None, + **kwargs + ): + super().__init__() # Initialize LLM base class + self.example_input_array = (torch.zeros(1, 1, dtype=torch.long),) + self.save_hyperparameters() + ### Config + self.actor_lr = actor_lr + self.critic_lr = critic_lr + self.discount_factor = discount_factor + self.tau = tau + + ### Manual Gradient Accumulation + self.accumulate_grad_batches = accumulate_grad_batches + self.automatic_optimization = False + + ### Initialization + self.actor = ActorModel( + get_device=lambda: self.device, model_name_or_path=model_name_or_path, actor_checkpoint=actor_checkpoint + ) + self.critic = RobertaCritic( + get_device=lambda: self.device, + discount_factor=discount_factor, + tau=tau, + expectile=critic_expectile, + from_checkpoint=critic_checkpoint, + ) + + self.actor_current_backward_step = 0 + self.critic_current_backward_step = 0 + self.critic_warmup_gradient_steps = 0 + + self.optimize_actor = lambda: ( + True + if self.critic_current_backward_step // self.accumulate_grad_batches + >= self.critic_warmup_gradient_steps + else False + ) + self.optimize_critic = lambda: optimize_critic + + def forward(self, observation, **kwargs): + action = self.actor.forward(observation, **kwargs) + return action + + def training_step(self, batch, batch_idx): + # if batch_idx == 3: + # return + + optimizer = self.optimizers() + mem = torch.cuda.memory_allocated() / torch.cuda.get_device_properties(0).total_memory + if mem > 0.8: + gc.collect() + torch.cuda.empty_cache() + + if self.optimize_critic(): + # scale losses by 1/N (for N batches of gradient accumulation) + critic_loss, critic_log = self.critic_loss(batch) + critic_loss /= self.accumulate_grad_batches + self.manual_backward(critic_loss) + self.critic_current_backward_step += 1 + self.log_dict(critic_log, sync_dist=True) + + # accumulate gradients of N batches + if self.critic_current_backward_step % self.accumulate_grad_batches == 0: + optimizer.step() + optimizer.zero_grad() + self.critic.soft_update_target_critic(self.tau) + + + if self.optimize_actor(): + # scale losses by 1/N (for N batches of gradient accumulation) + + actor_loss, actor_log = self.actor_loss(batch) + actor_loss /= self.accumulate_grad_batches + self.manual_backward(actor_loss) + self.actor_current_backward_step += 1 + self.log_dict(actor_log, sync_dist=True) + + # accumulate gradients of N batches + if self.actor_current_backward_step % self.accumulate_grad_batches == 0: + optimizer.step() + optimizer.zero_grad() + + def get_actor_log(self, loss, advantages, log_prob): + return { + "actor/loss": loss.detach(), + "actor/advantages.mean": advantages.detach().mean(), + "actor/advantages.max": torch.max(advantages.detach()), + "actor/advantages.min": torch.min(advantages.detach()), + "actor/log_prob.mean": torch.mean(log_prob.detach()), + "actor/log_prob.max": torch.max(log_prob.detach()), + "actor/log_prob.min": torch.min(log_prob.detach()), + } + + def configure_optimizers(self): + from torch.optim import Adam + + optimizer_params = [] + + if hasattr(self, 'actor') and hasattr(self.actor, 'parameters'): + optimizer_params.append({ + "params": self.actor.parameters(), + "lr": self.actor_lr + }) + + if self.optimize_critic and hasattr(self, 'critic') and hasattr(self.critic, 'parameters'): + optimizer_params.append({ + "params": self.critic.parameters(), + "lr": self.critic_lr + }) + + if not optimizer_params: + return None + + optimizer = Adam(optimizer_params) + return optimizer + + +class OfflineArcher(ActorCritic): + def __init__( + self, + model_name_or_path: str, + inv_temp: float, + actor_lr: float, + critic_lr: float, + tau: float, + accumulate_grad_batches: int, + discount_factor: float, + critic_expectile: float, + optimize_critic: bool, + actor_checkpoint: Optional[str] = None, + critic_checkpoint: Optional[str] = None, + **kwargs + ): + super().__init__( + model_name_or_path=model_name_or_path, + actor_lr=actor_lr, + critic_lr=critic_lr, + tau=tau, + accumulate_grad_batches=accumulate_grad_batches, + discount_factor=discount_factor, + critic_expectile=critic_expectile, + optimize_critic=optimize_critic, + actor_checkpoint=actor_checkpoint, + critic_checkpoint=critic_checkpoint, + **kwargs + ) + + self.inv_temp = inv_temp + + self.actor_loss = lambda batch: self.awr_loss(**batch) + self.critic_loss = lambda batch: self.critic.iql_loss(**batch) + + def awr_loss(self, observation, action, **kwargs): + log_prob = self.actor.get_logsum_prob(observation, action) + with torch.no_grad(): + advantages = self.critic.get_advantages(observation, action) + + advantages = advantages.flatten() + log_prob = log_prob.flatten() + factor = torch.exp(self.inv_temp * advantages) + loss_batch = -factor * log_prob + loss = loss_batch.mean() + + # ### Log and print what's happening + log = self.get_actor_log(loss=loss, advantages=advantages, log_prob=log_prob) + log = { + **log, + **{ + "actor/factor.mean": factor.detach().mean(), + "actor/factor.max": torch.max(factor.detach()), + "actor/factor.min": torch.min(factor.detach()), + }, + } + + return loss, log + + def configure_optimizers(self): + # 直接调用基类方法 + return super().configure_optimizers() + + @rank_zero_only + def on_save_checkpoint(self, checkpoint): + """保存 LoRA 适配器权重""" + super().on_save_checkpoint(checkpoint) + + save_dir = self.trainer.default_root_dir + os.makedirs(save_dir, exist_ok=True) + + # 保存 LoRA 适配器 + if hasattr(self.actor.model, "save_pretrained"): + self.actor.model.save_pretrained(save_dir) + + # 保存 tokenizer + if hasattr(self.actor, "tokenizer"): + self.actor.tokenizer.save_pretrained(save_dir) + + print(f"✅ LoRA adapter saved to: {save_dir}") + self.merge_and_save_lora(os.path.join(save_dir, "merged_model")) + + def merge_and_save_lora(self, save_dir): + """ + Merge the LoRA adapter weights into the base model and save the merged model and tokenizer. + """ + # Only proceed if the actor model has the correct method + try: + # 确保模型在CPU上且处于eval模式 + original_device = next(self.actor.model.parameters()).device + self.actor.model.to('cpu') + self.actor.model.eval() + + if hasattr(self.actor.model, "merge_and_unload"): + # 执行合并 + merged_model = self.actor.model.merge_and_unload() + + # 检查合并结果 + from peft import PeftModel + if isinstance(merged_model, PeftModel): + print(">>> [Warning] Still a PeftModel after merge. Using base_model.model...") + merged_model = merged_model.base_model.model + + # 保存合并后的模型 + merged_model.save_pretrained(os.path.join(save_dir, "merged_model")) + print(f"✅ Merged model saved to: {os.path.join(save_dir, 'merged_model')}") + else: + print("❌ merge_and_unload method not found in actor.model. Cannot merge LoRA weights.") + except Exception as e: + print(f"❌ Error merging LoRA weights: {e}") + import traceback + traceback.print_exc() + finally: + # 恢复原始设备 + self.actor.model.to(original_device) + + +class IQLKL(ActorCritic): + def __init__(self, kl_coeff: float, reference_actor_path, **kwargs): + super().__init__(**kwargs) + + self.kl_coeff = kl_coeff + self.reference_actor = ActorModel( + get_device=lambda: self.device, from_checkpoint=reference_actor_path + ) + + self.actor_loss = lambda batch: self.advantage_kl_loss(**batch) + self.critic_loss = lambda batch: self.critic.iql_loss(**batch) + + def advantage_kl_loss(self, observation, **kwargs): + reinforce_loss, generated_output = self.reinforce_loss(observation=observation) + with torch.no_grad(): + reference_log_prob = self.reference_actor.get_logsum_prob( + observation, generated_output["action"] + ) + + ratio = generated_output["log_prob"] - reference_log_prob + kl_loss = (ratio.detach() + 1.0) * generated_output["log_prob"] + loss = (1 - self.kl_coeff) * reinforce_loss + self.kl_coeff * kl_loss + log = generated_output["log"] + log = { + **log, + "reference_log_prob.mean": reference_log_prob.mean(), + "reference_log_prob.max": reference_log_prob.max(), + "reference_log_prob.min": reference_log_prob.min(), + } + log = { + **log, + "kl_loss.mean": kl_loss.mean(), + "kl_loss.max": kl_loss.max(), + "kl_loss.min": kl_loss.min(), + } + log = {**log, "actor_loss.mean": loss.mean(), "ratio": ratio.mean()} + + return loss.mean(), log + + def reinforce_loss(self, observation, **kwargs): + ### Reinforce Loss + action = self.actor.forward(observation) + log_prob = self.actor.get_logsum_prob(observation, action) + + with torch.no_grad(): + advantages = self.critic.get_advantages(observation, action) + + loss = -advantages.flatten() * log_prob + + ### Logging + log = self.get_actor_log( + loss=torch.mean(loss.detach()), advantages=advantages, log_prob=log_prob + ) + # self.log_dict(log) + return loss, { + "log_prob": log_prob, + "advantages": advantages, + "action": action, + "log": log, + } diff --git a/ArcherCritic.py b/ArcherCritic.py new file mode 100644 index 0000000000000000000000000000000000000000..a19311c511b84a58146cd94461ee5ebb6044734e --- /dev/null +++ b/ArcherCritic.py @@ -0,0 +1,161 @@ +import torch +from transformers import AutoModelForCausalLM, AutoTokenizer +from transformers import AutoTokenizer, RobertaModel +import torch.nn as nn +from transformers import RobertaTokenizer, RobertaModel +import logging +# A logger for this file +log = logging.getLogger(__name__) + +class ArcherDoubleCritic(torch.nn.Module): + def __init__(self, in_dim, out_dim): + super(ArcherDoubleCritic, self).__init__() + self.base_lm = RobertaModel.from_pretrained('roberta-base', torch_dtype=torch.float16) + + ################ + print("*** Master Warning - Are these used? *** ") + # self.base_lm.pooler.dense.weight = None + # self.base_lm.pooler.dense.bias = None + ############### + self.base_tokenizer = RobertaTokenizer.from_pretrained('roberta-base') + self.base_tokenizer.truncation_side = 'left' + self.critic1 = nn.Sequential(nn.Linear(in_dim*2, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + self.critic2 = nn.Sequential(nn.Linear(in_dim*2, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + self.v_critic1 = nn.Sequential(nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + self.v_critic2 = nn.Sequential(nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + + def get_q(self, observation, action, detach_model=False): + state_actions = [o + a for o,a in zip(observation, action)] + obs_ids = self.base_tokenizer(observation, padding = True, return_tensors='pt', truncation=True, max_length=512).to(self.base_lm.device) + if detach_model: + with torch.no_grad(): + lm_states = self.base_lm(**obs_ids).last_hidden_state[:,0] + else: + lm_states = self.base_lm(**obs_ids).last_hidden_state[:,0] + action_ids = self.base_tokenizer(action, padding = True, return_tensors='pt', truncation=True, max_length=512).to(self.base_lm.device) + if detach_model: + with torch.no_grad(): + action_states = self.base_lm(**action_ids).last_hidden_state[:,0] + else: + action_states = self.base_lm(**action_ids).last_hidden_state[:,0] + lm_states = torch.cat([lm_states, action_states], dim = 1) + return self.critic1(lm_states), self.critic2(lm_states) + + def get_v(self, observation,detach_model=False): + obs_ids = self.base_tokenizer(observation, padding = True, return_tensors='pt', truncation=True, max_length=512).to(self.base_lm.device) + if detach_model: + with torch.no_grad(): + lm_states = self.base_lm(**obs_ids).last_hidden_state[:,0] + else: + lm_states = self.base_lm(**obs_ids).last_hidden_state[:,0] + # print(action.size()) + return self.v_critic1(lm_states), self.v_critic2(lm_states) + + +class ArcherCritic(torch.nn.Module): + def __init__(self, in_dim=768, out_dim=4096, dropout = 0.5): + super(ArcherCritic, self).__init__() + self.model = AutoModelForCausalLM.from_pretrained('gpt2') + self.critic = ArcherDoubleCritic(in_dim = 768, out_dim = 1) + self.target_critic = ArcherDoubleCritic(in_dim = 768, out_dim = 1) + self.soft_update_target_critic(1) + self.tokenizer = AutoTokenizer.from_pretrained('gpt2', trust_remote_code=True) + self.tokenizer.truncation_side = 'left' + self.tokenizer.pad_token = self.tokenizer.eos_token + self.tokenizer.pad_token_id = self.tokenizer.eos_token_id + # self.device = device + self.dropout = torch.nn.Dropout(p=dropout) + self.softmax = torch.nn.Softmax(dim= -1) + + def get_action(self, observation): + obs_ids = self.tokenizer(observation, return_tensors='pt', padding=True, max_length=512).to(self.model.device) + obs_embeds = self.model.get_input_embeddings()(obs_ids["input_ids"]) + outputs = self.model.generate(inputs_embeds=obs_embeds, attention_mask=obs_ids['attention_mask'],\ + max_new_tokens=32, do_sample=True, \ + pad_token_id = self.tokenizer.eos_token_id)#.cpu() + raw_action = self.tokenizer.batch_decode(outputs, skip_special_tokens = True) + return raw_action + + def get_q(self, observation, action, detach_model=False): + return self.critic.get_q(observation, action, detach_model = detach_model) + + def get_v(self, inputs, detach_model=False): + return self.critic.get_v(inputs, detach_model = detach_model) + + def get_target_q(self, observation, action, detach_model=False): + return self.target_critic.get_q(observation, action, detach_model = detach_model) + + def get_log_prob(self, observation, action): + obs_ids = self.tokenizer(observation, return_tensors='pt', padding=True, max_length=512).to(self.model.device) + action_ids = self.tokenizer(action, return_tensors='pt', padding=True, max_length=512).to(self.model.device) + action_embeds = self.model.get_input_embeddings()(action_ids["input_ids"]).detach() + obs_embeds = self.model.get_input_embeddings()(obs_ids["input_ids"]).detach() + input_embeds = torch.cat([obs_embeds, action_embeds], dim = 1) + attention_mask = torch.cat([obs_ids["attention_mask"], action_ids["attention_mask"]],\ + dim = 1) + outputs = self.model(inputs_embeds = input_embeds, attention_mask = attention_mask) + prediction_probs = self.softmax(outputs.logits) + selected_prediction_probs = torch.take_along_dim(prediction_probs[:, obs_ids["attention_mask"].size(1)-1:-1],\ + action_ids["input_ids"].unsqueeze(2), dim=2).squeeze(2) + logsum_probs = torch.sum(torch.log(selected_prediction_probs)*action_ids["attention_mask"], dim = 1) + return logsum_probs + + def soft_update_target_critic(self, tau): + # for target_critic, critic in zip(self.target_critics, self.critics): + for target_param, param in zip( + self.target_critic.parameters(), self.critic.parameters() + ): + target_param.data.copy_( + target_param.data * (1.0 - tau) + param.data * tau + ) + +class DoubleCritic(torch.nn.Module): + """ + a double critic without base lm + """ + def __init__(self, in_dim, out_dim): + super(DoubleCritic, self).__init__() + # self.device = device + self.critic1 = nn.Sequential(nn.Linear(in_dim*2, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + self.critic2 = nn.Sequential(nn.Linear(in_dim*2, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + self.v_critic1 = nn.Sequential(nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + self.v_critic2 = nn.Sequential(nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, in_dim),\ + nn.ReLU(),\ + nn.Linear(in_dim, out_dim))#.to(device) + + def get_q(self, observation, action, detach_model=False): + lm_states = torch.cat([observation, action], dim = 1) + return self.critic1(lm_states), self.critic2(lm_states) + + def get_v(self, observation,detach_model=False): + return self.v_critic1(observation), self.v_critic2(observation) diff --git a/Dataset.py b/Dataset.py new file mode 100644 index 0000000000000000000000000000000000000000..dc27c48d45211123ef12419b6038bc26b0958cba --- /dev/null +++ b/Dataset.py @@ -0,0 +1,328 @@ +from torch.utils.data import Dataset +import numpy as np +import copy +from torch.utils.data import Dataset + +class DummyDataset(Dataset): + def __init__(self, buffer): + self.buffer = buffer + + def __len__(self): + return len(self.buffer) + + def __getitem__(self, idx): + return self.buffer[idx] + + +class Transition: + def __init__(self, observation, action, reward, next_observation, done, **kwargs): + self.observation = observation + self.action = action + self.reward = np.single(reward) + self.next_observation = next_observation + + if isinstance(done, bool): + self.done = done + elif done == 'False': + self.done = False + elif done == 'True': + self.done = True + else: + raise ValueError + + # internal, to see how many times a certain transition was sampled + self.times_was_sampled = 0 + def as_dict(self, as_string = False): + return { + "observation": self.observation, + "action": self.action, + "reward": self.reward if as_string == False else str(self.reward), + "next_observation": self.next_observation, + "done": self.done if as_string == False else str(self.done) + } + def __str__(self): + printout = '\n' + for key in self.as_dict(): + printout += "\n" + key + ':' + printout += '\n' + str(self.as_dict()[key]) + return printout + +class Trajectory: + def __init__(self): + self.transitions = [] + self.info = {} + def __len__(self): + return len(self.transitions) + def check_consistency(self): + assert(any([transition.done for transition in self.transitions[:-1]]) == False) # should not be done until the end + assert(self.transitions[-1].done == True ) + for t in range(1,len(self.transitions)): + prior_transition = self.transitions[t-1] + current_transition = self.transitions[t] + assert(prior_transition.next_observation == current_transition.observation) + def get_rewards(self): + return [transition.reward for transition in self.transitions] + def get_return(self): + return sum([transition.reward for transition in self.transitions]) + def append(self, transition): + assert(self.transitions == [] or self.transitions[-1].done == False) + self.transitions.append(Transition(**transition)) + def __str__(self): + printout = '\n*** Trajectory Begins *** \n' + printout += "\nTrajectory Length: " + str(len(self)) + for idx, transition in enumerate(self.transitions): + printout += "\nTransition: " + str(idx) + printout += "\n" + transition.__str__() + if self.info != None: + printout += "\nFound Special Items" + printout += str(self.info) + printout += '\n *** Trajectory Ends **** \n' + return printout + +class TrajectoryDataset(Dataset): + def __init__(self): + self.trajectories = [] + self.samples = [] # pointer list for fast sampling + self._last_oar = None # Last (observation, action, reward) for sequential addition + def __len__(self): + return len(self.samples) + def __getitem__(self, idx): + self.samples[idx].times_was_sampled += 1 + ### Must return a copy to avoid issues if further processing is done + return copy.deepcopy(self.samples[idx].as_dict()) + def append_trajectory(self, trajectory: Trajectory): + trajectory.check_consistency() + assert(self.last_trajectory_reached_end()) + for transition in trajectory.transitions: + self.append_sample_sequentially(copy.deepcopy(transition.as_dict())) + self.trajectories[-1].info = copy.deepcopy(trajectory.info) + self.trajectories[-1].check_consistency() + def append_observation_action_reward(self, observation, action, reward): + if self._last_oar != None: + self.append_sample_sequentially({"observation": self._last_oar["observation"], + "action": self._last_oar["action"], + "reward": self._last_oar["reward"], + "next_observation": observation, + "done": False }) + self._last_oar = {"observation": observation, + "action": action, + "reward": reward} + def append_terminal_observation(self, observation, trajectory_info = None): + assert self._last_oar != None + self.append_sample_sequentially({"observation": self._last_oar["observation"], + "action": self._last_oar["action"], + "reward": self._last_oar["reward"], + "next_observation": observation, + "done": True }) + self._last_oar = None + if trajectory_info != None: + self.trajectories[-1].info = trajectory_info + self.trajectories[-1].check_consistency() + + def last_trajectory_reached_end(self): + return (self.trajectories == [] or self.trajectories[-1].transitions[-1].done) + + def append_sample_sequentially(self, transition): + ### is the trajectory new? + if self.last_trajectory_reached_end(): + self.trajectories.append(Trajectory()) + self.trajectories[-1].transitions.append(Transition(**transition)) + self.samples.append(self.trajectories[-1].transitions[-1]) + def nTrajectories(self): + return len(self.trajectories) + def get_all_trajectory_returns(self): + return np.asarray([trajectory.get_return() for trajectory in self.trajectories]) + def check_consistency(self): + assert (sum([len(trajectory) for trajectory in self.trajectories]) == len(self.samples)) + for trajectory in self.trajectories: + trajectory.check_consistency() + def sample(self, batch_size=None): + if batch_size is None: + batch_size = self.batch_size + rand_indices = np.random.randint(0, len(self.samples), size=(batch_size,)) + # rand_indices = [np.random.randint(0, len(self.samples)) for _ in range(batch_size)] + for idx in rand_indices: + self.samples[idx].times_was_sampled += 1 + return { + "observation": [self.samples[idx].observation for idx in rand_indices], + "action": [self.samples[idx].action for idx in rand_indices], + "reward": [self.samples[idx].reward for idx in rand_indices], + "next_observation": [self.samples[idx].next_observation for idx in rand_indices], + "done": [self.samples[idx].done for idx in rand_indices], + } + def mean_trajectory_return(self): + return np.mean(self.get_all_trajectory_returns()) + def std_trajectory_return(self): + return np.std(self.get_all_trajectory_returns()) + def merge(self, dataset): + self.check_consistency() + dataset.check_consistency() + for trajectory in dataset.trajectories: + for transition in trajectory.transitions: + self.append_sample_sequentially(transition.as_dict()) + self.trajectories[-1].info = copy.deepcopy(trajectory.info) + self.check_consistency() + # assert(self.batch_size == dataset.batch_size) + def __str__(self): + printout = '\n \n ' + printout += '\n ************************ ' + printout += '\n *** Printing Dataset *** ' + printout += '\n ************************ ' + printout += '\n \n ' + printout += '\n Number of Samples : ' + str(len(self)) + printout += '\n Dataset Trajectories : ' + str(self.nTrajectories()) + '\n' + for idx, trajectory in enumerate(self.trajectories): + printout += "\n >>> Trajectory id: " + str(idx) + '\n' + printout += trajectory.__str__() + if self._last_oar != None: + printout += "\n !!! Found incomplete transition !!! \n" + for key in self._last_oar: + printout += key + '\n' + printout += str(self._last_oar[key]) + "\n" + printout += '\n ************************ ' + printout += '\n *** Dataset Printed *** ' + printout += '\n ************************ ' + return printout + + def keep_top_fraction_of_trajectories(self, fraction: float, from_high_to_low = True): + self.sort(from_high_to_low=from_high_to_low) + trajectories = self.trajectories + import math + nTraj_to_keep = int(fraction * self.nTrajectories()) + self.__init__() + for i in range(nTraj_to_keep): + self.append_trajectory(trajectories[i]) + print("*** Kept ", self.nTrajectories(), " trajectories") + + def keep_bottom_fraction_of_trajectories(self, fraction: float): + self.keep_top_fraction_of_trajectories(fraction=fraction, from_high_to_low=False) + + + def max_trajectory_return(self): + return max(self.get_all_trajectory_returns()) + + def argmax_trajectory_return(self): + return np.argmax(self.get_all_trajectory_returns()) + + def min_trajectory_return(self): + return min(self.get_all_trajectory_returns()) + + def argmin_trajectory_return(self): + return np.argmin(self.get_all_trajectory_returns()) + + def sort(self, from_high_to_low): + print("Warning: new dataset created!") + returns = [trajectory.get_return() for trajectory in self.trajectories] + sorted_trajectories = sort_list(self.trajectories, returns, from_high_to_low) + self.__init__() + for traj in sorted_trajectories: + self.append_trajectory(traj) + + # useful to set all rewards to eg -1 and encourage reaching the goal faster + def set_all_rewards_to_value(self, value): + for sample in self.samples: + sample.reward = np.single(value) + + def scale_all_rewards_by_value(self, value): + for sample in self.samples: + sample.reward *= np.single(value) + + def add_value_to_all_rewards(self, value): + for sample in self.samples: + sample.reward += np.single(value) + + def increase_final_reward_by_value(self, value): + for trajectory in self.trajectories: + trajectory.transitions[-1].reward += np.single(value) + + def append_eos_token_to_all_actions(self, eos_token): + for sample in self.samples: + sample.action += eos_token + + def push_all_rewards_at_the_end_of_the_trajectory(self): + for trajectory in self.trajectories: + trajectory.transitions[-1].reward = np.single(trajectory.get_return()) + for transition in trajectory.transitions[:-1]: + transition.reward = np.single(0) + assert(- len(trajectory) == trajectory.get_return() == trajectory.transitions[-1].reward) + + def save(self, filename): + import json + self.check_consistency() + with open(filename, "w") as final: + json.dump([sample.as_dict(as_string = True) for sample in self.samples], final) + + def load(self, filename): + import json + with open(filename, "r") as final: + data = json.load(final) + for sample in data: + self.append_sample_sequentially(sample) + + def times_was_sampled(self): + return [sample.times_was_sampled for sample in self.samples] + + def keep_only_trajectories_with_exact_key_and_value(self, key, value): + trajectories = self.trajectories + new_dataset = TrajectoryDataset() + for trajectory in trajectories: + if trajectory.info[key] == value: + new_dataset.append_trajectory(trajectory) + return new_dataset + + def construct_tabular_state_action_space(self): + self.state_space = Counter() + self.action_space = Counter() + self.state_action_space = Counter() + for sample in self.samples: + self.state_space.add(sample.observation) + self.action_space.add(sample.action) + self.state_action_space.add((sample.observation, sample.action)) + + def assert_deterministic(self): + successor_states = {} + rewards = {} + for sample in self.samples: + sa = (sample.observation, sample.action) + + if sa not in rewards: + rewards[sa] = sample.reward + + else: + assert(rewards[sa] == sample.reward) + + if sample.done: # end transition may be ill-defined + continue + + if sa not in successor_states: + successor_states[sa] = sample.next_observation + else: + assert(successor_states[sa] == sample.next_observation) + +class Counter(): + def __init__(self): + self.register = {} + def add(self, item): + if item not in self.register: + self.register[item] = 1 + else: + self.register[item] += 1 + def contains(self, item): + return item in self.register + + def n_samples(self, item): + return self.register[item] + +class EmptyDataset(): + def __init__(self, length): + self.length = length + + def __len__(self): + return self.length + + def __getitem__(self, idx): + return [0] + +def sort_list(list1, list2, from_high_to_low): + # Sorting the List1 based on List2 + return [val for (_, val) in sorted(zip(list2, list1), key=lambda x: x[0], reverse=from_high_to_low)] diff --git a/Llama3-8B-I_Strategic_log.txt b/Llama3-8B-I_Strategic_log.txt new file mode 100644 index 0000000000000000000000000000000000000000..251a3695f8fb438ae7182ec25b4ad2f97889dcc1 --- /dev/null +++ b/Llama3-8B-I_Strategic_log.txt @@ -0,0 +1,233 @@ +The following values were not passed to `accelerate launch` and had defaults used instead: + `--num_cpu_threads_per_process` was set to `8` to improve out-of-box performance when training on CPUs +To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`. +[2025-09-16 00:15:01,530] [INFO] [real_accelerator.py:260:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-09-16 00:15:03,007] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +W0916 00:15:03.145761 2266983 site-packages/torch/distributed/run.py:774] +W0916 00:15:03.145761 2266983 site-packages/torch/distributed/run.py:774] ***************************************** +W0916 00:15:03.145761 2266983 site-packages/torch/distributed/run.py:774] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. +W0916 00:15:03.145761 2266983 site-packages/torch/distributed/run.py:774] ***************************************** +[rank: 3] Seed set to 42 +[rank: 5] Seed set to 42 +[rank: 4] Seed set to 42 +[rank: 0] Seed set to 42 +[rank: 6] Seed set to 42 +[rank: 7] Seed set to 42 +[rank: 2] Seed set to 42 +[rank: 1] Seed set to 42 +[rank: 3] Seed set to 42 +`torch_dtype` is deprecated! Use `dtype` instead! + Loading checkpoint shards: 0%| | 0/4 [00:00 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 5%|▌ | 97/1791 [25:55<7:32:40, 0.06it/s, v_num=bc8f] Epoch 0: 5%|▌ | 97/1791 [25:55<7:32:40, 0.06it/s, v_num=bc8f] Epoch 0: 5%|▌ | 98/1791 [26:11<7:32:23, 0.06it/s, v_num=bc8f] Epoch 0: 5%|▌ | 98/1791 [26:11<7:32:23, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 99/1791 [26:26<7:31:55, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 99/1791 [26:26<7:31:55, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 100/1791 [26:42<7:31:30, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 100/1791 [26:42<7:31:30, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 101/1791 [26:57<7:31:03, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 101/1791 [26:57<7:31:03, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 102/1791 [27:12<7:30:36, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 102/1791 [27:12<7:30:36, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 103/1791 [27:27<7:30:06, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 103/1791 [27:27<7:30:06, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 104/1791 [27:44<7:30:01, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 104/1791 [27:44<7:30:01, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 105/1791 [28:00<7:29:36, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 105/1791 [28:00<7:29:36, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 106/1791 [28:15<7:29:13, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 106/1791 [28:15<7:29:13, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 107/1791 [28:31<7:28:51, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 107/1791 [28:31<7:28:51, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 108/1791 [28:46<7:28:27, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 108/1791 [28:46<7:28:27, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 109/1791 [29:02<7:28:11, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 109/1791 [29:02<7:28:11, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 110/1791 [29:18<7:27:54, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 110/1791 [29:18<7:27:54, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 111/1791 [29:34<7:27:38, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▌ | 111/1791 [29:34<7:27:38, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 112/1791 [29:51<7:27:30, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 112/1791 [29:51<7:27:30, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 113/1791 [30:06<7:27:07, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 113/1791 [30:06<7:27:07, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 114/1791 [30:22<7:26:50, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 114/1791 [30:22<7:26:50, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 115/1791 [30:38<7:26:36, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 115/1791 [30:38<7:26:37, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 116/1791 [30:54<7:26:12, 0.06it/s, v_num=bc8f] Epoch 0: 6%|▋ | 116/1791 [30:54<7:26:12, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 117/1791 [31:09<7:25:48, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 117/1791 [31:09<7:25:48, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 118/1791 [31:25<7:25:27, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 118/1791 [31:25<7:25:27, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 119/1791 [31:40<7:25:01, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 119/1791 [31:40<7:25:01, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 120/1791 [31:57<7:25:00, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 120/1791 [31:57<7:25:00, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 121/1791 [32:13<7:24:40, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 121/1791 [32:13<7:24:40, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 122/1791 [32:28<7:24:15, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 122/1791 [32:28<7:24:15, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 123/1791 [32:44<7:24:03, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 123/1791 [32:44<7:24:03, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 124/1791 [33:00<7:23:44, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 124/1791 [33:00<7:23:44, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 125/1791 [33:16<7:23:27, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 125/1791 [33:16<7:23:27, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 126/1791 [33:32<7:23:07, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 126/1791 [33:32<7:23:07, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 127/1791 [33:46<7:22:31, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 127/1791 [33:46<7:22:31, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 128/1791 [34:03<7:22:30, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 128/1791 [34:03<7:22:30, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 129/1791 [34:19<7:22:14, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 129/1791 [34:19<7:22:14, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 130/1791 [34:35<7:21:55, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 130/1791 [34:35<7:21:55, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 131/1791 [34:51<7:21:37, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 131/1791 [34:51<7:21:37, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 132/1791 [35:06<7:21:15, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 132/1791 [35:06<7:21:15, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 133/1791 [35:21<7:20:47, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 133/1791 [35:21<7:20:47, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 134/1791 [35:37<7:20:29, 0.06it/s, v_num=bc8f] Epoch 0: 7%|▋ | 134/1791 [35:37<7:20:29, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 135/1791 [35:52<7:20:00, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 135/1791 [35:52<7:20:00, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 136/1791 [36:09<7:19:58, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 136/1791 [36:09<7:19:58, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 137/1791 [36:24<7:19:34, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 137/1791 [36:24<7:19:34, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 138/1791 [36:40<7:19:13, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 138/1791 [36:40<7:19:13, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 139/1791 [36:55<7:18:53, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 139/1791 [36:55<7:18:53, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 140/1791 [37:10<7:18:29, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 140/1791 [37:10<7:18:29, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 141/1791 [37:26<7:18:11, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 141/1791 [37:26<7:18:12, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 142/1791 [37:42<7:17:54, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 142/1791 [37:42<7:17:54, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 143/1791 [37:57<7:17:29, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 143/1791 [37:57<7:17:29, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 144/1791 [38:14<7:17:27, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 144/1791 [38:14<7:17:27, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 145/1791 [38:30<7:17:05, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 145/1791 [38:30<7:17:05, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 146/1791 [38:45<7:16:46, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 146/1791 [38:45<7:16:46, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 147/1791 [39:01<7:16:26, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 147/1791 [39:01<7:16:26, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 148/1791 [39:16<7:16:00, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 148/1791 [39:16<7:16:00, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 149/1791 [39:31<7:15:38, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 149/1791 [39:31<7:15:38, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 150/1791 [39:46<7:15:10, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 150/1791 [39:46<7:15:10, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 151/1791 [40:02<7:14:52, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 151/1791 [40:02<7:14:52, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 152/1791 [40:18<7:14:38, 0.06it/s, v_num=bc8f] Epoch 0: 8%|▊ | 152/1791 [40:18<7:14:38, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 153/1791 [40:33<7:14:10, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 153/1791 [40:33<7:14:10, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 154/1791 [40:48<7:13:46, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 154/1791 [40:48<7:13:46, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 155/1791 [41:03<7:13:24, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 155/1791 [41:03<7:13:24, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 156/1791 [41:19<7:13:05, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▊ | 156/1791 [41:19<7:13:05, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 157/1791 [41:34<7:12:44, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 157/1791 [41:34<7:12:44, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 158/1791 [41:50<7:12:22, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 158/1791 [41:50<7:12:22, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 159/1791 [42:06<7:12:07, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 159/1791 [42:06<7:12:07, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 160/1791 [42:23<7:12:09, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 160/1791 [42:23<7:12:09, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 161/1791 [42:39<7:11:49, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 161/1791 [42:39<7:11:49, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 162/1791 [42:55<7:11:36, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 162/1791 [42:55<7:11:37, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 163/1791 [43:11<7:11:22, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 163/1791 [43:11<7:11:22, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 164/1791 [43:26<7:11:02, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 164/1791 [43:26<7:11:02, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 165/1791 [43:42<7:10:46, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 165/1791 [43:42<7:10:46, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 166/1791 [43:58<7:10:26, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 166/1791 [43:58<7:10:26, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 167/1791 [44:14<7:10:10, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 167/1791 [44:14<7:10:10, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 168/1791 [44:31<7:10:04, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 168/1791 [44:31<7:10:04, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 169/1791 [44:46<7:09:44, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 169/1791 [44:46<7:09:44, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 170/1791 [45:02<7:09:26, 0.06it/s, v_num=bc8f] Epoch 0: 9%|▉ | 170/1791 [45:02<7:09:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 171/1791 [45:17<7:09:08, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 171/1791 [45:17<7:09:08, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 172/1791 [45:34<7:08:55, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 172/1791 [45:34<7:08:55, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 173/1791 [45:50<7:08:39, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 173/1791 [45:50<7:08:39, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 174/1791 [46:06<7:08:30, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 174/1791 [46:06<7:08:30, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 175/1791 [46:22<7:08:11, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 175/1791 [46:22<7:08:11, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 176/1791 [46:39<7:08:09, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 176/1791 [46:39<7:08:09, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 177/1791 [46:55<7:07:49, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 177/1791 [46:55<7:07:49, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 178/1791 [47:10<7:07:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 178/1791 [47:10<7:07:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 179/1791 [47:25<7:07:05, 0.06it/s, v_num=bc8f] Epoch 0: 10%|▉ | 179/1791 [47:25<7:07:05, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 180/1791 [47:40<7:06:45, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 180/1791 [47:40<7:06:45, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 181/1791 [47:56<7:06:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 181/1791 [47:56<7:06:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 182/1791 [48:12<7:06:09, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 182/1791 [48:12<7:06:09, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 183/1791 [48:27<7:05:51, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 183/1791 [48:27<7:05:51, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 184/1791 [48:45<7:05:47, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 184/1791 [48:45<7:05:47, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 185/1791 [49:00<7:05:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 185/1791 [49:00<7:05:26, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 186/1791 [49:16<7:05:11, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 186/1791 [49:16<7:05:11, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 187/1791 [49:32<7:04:55, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 187/1791 [49:32<7:04:55, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 188/1791 [49:48<7:04:38, 0.06it/s, v_num=bc8f] Epoch 0: 10%|█ | 188/1791 [49:48<7:04:38, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 189/1791 [50:03<7:04:15, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 189/1791 [50:03<7:04:15, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 190/1791 [50:19<7:03:59, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 190/1791 [50:19<7:03:59, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 191/1791 [50:34<7:03:39, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 191/1791 [50:34<7:03:39, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 192/1791 [50:51<7:03:32, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 192/1791 [50:51<7:03:32, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 193/1791 [51:07<7:03:15, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 193/1791 [51:07<7:03:15, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 194/1791 [51:22<7:02:58, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 194/1791 [51:22<7:02:59, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 195/1791 [51:38<7:02:39, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 195/1791 [51:38<7:02:39, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 196/1791 [51:54<7:02:22, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 196/1791 [51:54<7:02:22, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 197/1791 [52:09<7:02:03, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 197/1791 [52:09<7:02:03, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 198/1791 [52:24<7:01:42, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 198/1791 [52:24<7:01:42, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 199/1791 [52:41<7:01:30, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 199/1791 [52:41<7:01:30, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 200/1791 [52:58<7:01:27, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 200/1791 [52:58<7:01:27, 0.06it/s, v_num=bc8f]Token indices sequence length is longer than the specified maximum sequence length for this model (1052 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 11%|█ | 201/1791 [53:15<7:01:13, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█ | 201/1791 [53:15<7:01:13, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 202/1791 [53:30<7:00:54, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 202/1791 [53:30<7:00:54, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 203/1791 [53:46<7:00:37, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 203/1791 [53:46<7:00:37, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 204/1791 [54:01<7:00:17, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 204/1791 [54:01<7:00:17, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 205/1791 [54:17<6:59:58, 0.06it/s, v_num=bc8f] Epoch 0: 11%|█▏ | 205/1791 [54:17<6:59:58, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 206/1791 [54:32<6:59:42, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 206/1791 [54:32<6:59:42, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 207/1791 [54:48<6:59:23, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 207/1791 [54:48<6:59:23, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 208/1791 [55:05<6:59:18, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 208/1791 [55:05<6:59:18, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 209/1791 [55:21<6:59:02, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 209/1791 [55:21<6:59:02, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 210/1791 [55:37<6:58:47, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 210/1791 [55:37<6:58:47, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 211/1791 [55:52<6:58:26, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 211/1791 [55:52<6:58:26, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 212/1791 [56:09<6:58:13, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 212/1791 [56:09<6:58:13, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 213/1791 [56:25<6:57:58, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 213/1791 [56:25<6:57:58, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 214/1791 [56:40<6:57:38, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 214/1791 [56:40<6:57:38, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 215/1791 [56:56<6:57:21, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 215/1791 [56:56<6:57:21, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 216/1791 [57:13<6:57:16, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 216/1791 [57:13<6:57:16, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 217/1791 [57:29<6:57:03, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 217/1791 [57:29<6:57:03, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 218/1791 [57:45<6:56:48, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 218/1791 [57:45<6:56:49, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 219/1791 [58:02<6:56:36, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 219/1791 [58:02<6:56:36, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 220/1791 [58:18<6:56:23, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 220/1791 [58:18<6:56:23, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 221/1791 [58:34<6:56:03, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 221/1791 [58:34<6:56:03, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 222/1791 [58:49<6:55:47, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 222/1791 [58:49<6:55:47, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 223/1791 [59:05<6:55:29, 0.06it/s, v_num=bc8f] Epoch 0: 12%|█▏ | 223/1791 [59:05<6:55:29, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 224/1791 [59:22<6:55:24, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 224/1791 [59:22<6:55:24, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 225/1791 [59:38<6:55:03, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 225/1791 [59:38<6:55:03, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 226/1791 [59:53<6:54:43, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 226/1791 [59:53<6:54:43, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 227/1791 [1:00:09<6:54:26, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 227/1791 [1:00:09<6:54:26, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 228/1791 [1:00:24<6:54:07, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 228/1791 [1:00:24<6:54:07, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 229/1791 [1:00:40<6:53:49, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 229/1791 [1:00:40<6:53:49, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 230/1791 [1:00:55<6:53:28, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 230/1791 [1:00:55<6:53:28, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 231/1791 [1:01:10<6:53:09, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 231/1791 [1:01:10<6:53:09, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 232/1791 [1:01:27<6:53:01, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 232/1791 [1:01:27<6:53:01, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 233/1791 [1:01:43<6:52:43, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 233/1791 [1:01:43<6:52:43, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 234/1791 [1:01:59<6:52:30, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 234/1791 [1:01:59<6:52:30, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 235/1791 [1:02:15<6:52:13, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 235/1791 [1:02:15<6:52:13, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 236/1791 [1:02:31<6:51:55, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 236/1791 [1:02:31<6:51:55, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 237/1791 [1:02:46<6:51:36, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 237/1791 [1:02:46<6:51:36, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 238/1791 [1:03:02<6:51:23, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 238/1791 [1:03:02<6:51:23, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 239/1791 [1:03:18<6:51:08, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 239/1791 [1:03:18<6:51:08, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 240/1791 [1:03:35<6:50:58, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 240/1791 [1:03:35<6:50:58, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 241/1791 [1:03:51<6:50:40, 0.06it/s, v_num=bc8f] Epoch 0: 13%|█▎ | 241/1791 [1:03:51<6:50:40, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 242/1791 [1:04:06<6:50:21, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 242/1791 [1:04:06<6:50:21, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 243/1791 [1:04:21<6:50:01, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 243/1791 [1:04:21<6:50:01, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 244/1791 [1:04:37<6:49:44, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 244/1791 [1:04:37<6:49:44, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 245/1791 [1:04:52<6:49:24, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 245/1791 [1:04:52<6:49:24, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 246/1791 [1:05:08<6:49:05, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▎ | 246/1791 [1:05:08<6:49:05, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 247/1791 [1:05:23<6:48:48, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 247/1791 [1:05:23<6:48:48, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 248/1791 [1:05:40<6:48:39, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 248/1791 [1:05:40<6:48:39, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 249/1791 [1:05:56<6:48:20, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 249/1791 [1:05:56<6:48:20, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 250/1791 [1:06:11<6:47:58, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 250/1791 [1:06:11<6:47:58, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 251/1791 [1:06:26<6:47:40, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 251/1791 [1:06:26<6:47:40, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 252/1791 [1:06:42<6:47:21, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 252/1791 [1:06:42<6:47:21, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 253/1791 [1:06:57<6:47:02, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 253/1791 [1:06:57<6:47:02, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 254/1791 [1:07:13<6:46:48, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 254/1791 [1:07:13<6:46:48, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 255/1791 [1:07:29<6:46:30, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 255/1791 [1:07:29<6:46:30, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 256/1791 [1:07:45<6:46:19, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 256/1791 [1:07:45<6:46:19, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 257/1791 [1:08:01<6:46:03, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 257/1791 [1:08:01<6:46:03, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 258/1791 [1:08:17<6:45:45, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 258/1791 [1:08:17<6:45:45, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 259/1791 [1:08:32<6:45:27, 0.06it/s, v_num=bc8f] Epoch 0: 14%|█▍ | 259/1791 [1:08:32<6:45:27, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 260/1791 [1:08:48<6:45:09, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 260/1791 [1:08:48<6:45:09, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 261/1791 [1:09:04<6:44:54, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 261/1791 [1:09:04<6:44:54, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 262/1791 [1:09:20<6:44:39, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 262/1791 [1:09:20<6:44:39, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 263/1791 [1:09:36<6:44:22, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 263/1791 [1:09:36<6:44:22, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 264/1791 [1:09:52<6:44:11, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 264/1791 [1:09:52<6:44:11, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 265/1791 [1:10:08<6:43:56, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 265/1791 [1:10:08<6:43:56, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 266/1791 [1:10:24<6:43:42, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 266/1791 [1:10:24<6:43:42, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 267/1791 [1:10:40<6:43:24, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 267/1791 [1:10:40<6:43:24, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 268/1791 [1:10:56<6:43:07, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▍ | 268/1791 [1:10:56<6:43:07, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 269/1791 [1:11:11<6:42:49, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 269/1791 [1:11:11<6:42:49, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 270/1791 [1:11:27<6:42:31, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 270/1791 [1:11:27<6:42:31, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 271/1791 [1:11:42<6:42:14, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 271/1791 [1:11:42<6:42:14, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 272/1791 [1:11:59<6:42:01, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 272/1791 [1:11:59<6:42:01, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 273/1791 [1:12:15<6:41:44, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 273/1791 [1:12:15<6:41:44, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 274/1791 [1:12:30<6:41:25, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 274/1791 [1:12:30<6:41:25, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 275/1791 [1:12:45<6:41:07, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 275/1791 [1:12:45<6:41:07, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 276/1791 [1:13:00<6:40:47, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 276/1791 [1:13:00<6:40:47, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 277/1791 [1:13:16<6:40:30, 0.06it/s, v_num=bc8f] Epoch 0: 15%|█▌ | 277/1791 [1:13:16<6:40:30, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 278/1791 [1:13:32<6:40:12, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 278/1791 [1:13:32<6:40:12, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 279/1791 [1:13:47<6:39:56, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 279/1791 [1:13:47<6:39:56, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 280/1791 [1:14:04<6:39:46, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 280/1791 [1:14:04<6:39:46, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 281/1791 [1:14:20<6:39:28, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 281/1791 [1:14:20<6:39:28, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 282/1791 [1:14:35<6:39:10, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 282/1791 [1:14:35<6:39:10, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 283/1791 [1:14:51<6:38:51, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 283/1791 [1:14:51<6:38:51, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 284/1791 [1:15:07<6:38:36, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 284/1791 [1:15:07<6:38:36, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 285/1791 [1:15:23<6:38:22, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 285/1791 [1:15:23<6:38:22, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 286/1791 [1:15:38<6:38:04, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 286/1791 [1:15:38<6:38:04, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 287/1791 [1:15:55<6:37:50, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 287/1791 [1:15:55<6:37:50, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 288/1791 [1:16:12<6:37:43, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 288/1791 [1:16:12<6:37:43, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 289/1791 [1:16:28<6:37:25, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 289/1791 [1:16:28<6:37:26, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 290/1791 [1:16:43<6:37:09, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 290/1791 [1:16:43<6:37:09, 0.06it/s, v_num=bc8f]Token indices sequence length is longer than the specified maximum sequence length for this model (1063 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 16%|█▌ | 291/1791 [1:16:59<6:36:52, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▌ | 291/1791 [1:16:59<6:36:52, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 292/1791 [1:17:15<6:36:37, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 292/1791 [1:17:15<6:36:37, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 293/1791 [1:17:31<6:36:19, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 293/1791 [1:17:31<6:36:19, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 294/1791 [1:17:46<6:36:02, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 294/1791 [1:17:46<6:36:02, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 295/1791 [1:18:02<6:35:44, 0.06it/s, v_num=bc8f] Epoch 0: 16%|█▋ | 295/1791 [1:18:02<6:35:44, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 296/1791 [1:18:19<6:35:33, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 296/1791 [1:18:19<6:35:33, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 297/1791 [1:18:34<6:35:17, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 297/1791 [1:18:34<6:35:17, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 298/1791 [1:18:50<6:35:00, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 298/1791 [1:18:50<6:35:00, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 299/1791 [1:19:06<6:34:42, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 299/1791 [1:19:06<6:34:42, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 300/1791 [1:19:21<6:34:24, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 300/1791 [1:19:21<6:34:24, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 301/1791 [1:19:36<6:34:06, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 301/1791 [1:19:36<6:34:06, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 302/1791 [1:19:52<6:33:48, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 302/1791 [1:19:52<6:33:48, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 303/1791 [1:20:07<6:33:31, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 303/1791 [1:20:07<6:33:31, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 304/1791 [1:20:25<6:33:21, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 304/1791 [1:20:25<6:33:21, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 305/1791 [1:20:41<6:33:06, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 305/1791 [1:20:41<6:33:06, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 306/1791 [1:20:57<6:32:51, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 306/1791 [1:20:57<6:32:51, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 307/1791 [1:21:13<6:32:35, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 307/1791 [1:21:13<6:32:35, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 308/1791 [1:21:29<6:32:20, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 308/1791 [1:21:29<6:32:20, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 309/1791 [1:21:44<6:32:04, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 309/1791 [1:21:44<6:32:04, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 310/1791 [1:22:00<6:31:48, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 310/1791 [1:22:00<6:31:48, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 311/1791 [1:22:16<6:31:32, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 311/1791 [1:22:16<6:31:32, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 312/1791 [1:22:33<6:31:23, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 312/1791 [1:22:33<6:31:23, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 313/1791 [1:22:49<6:31:05, 0.06it/s, v_num=bc8f] Epoch 0: 17%|█▋ | 313/1791 [1:22:49<6:31:05, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 314/1791 [1:23:05<6:30:49, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 314/1791 [1:23:05<6:30:50, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 315/1791 [1:23:21<6:30:34, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 315/1791 [1:23:21<6:30:34, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 316/1791 [1:23:36<6:30:17, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 316/1791 [1:23:36<6:30:17, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 317/1791 [1:23:52<6:30:01, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 317/1791 [1:23:52<6:30:01, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 318/1791 [1:24:08<6:29:43, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 318/1791 [1:24:08<6:29:43, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 319/1791 [1:24:23<6:29:26, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 319/1791 [1:24:23<6:29:26, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 320/1791 [1:24:40<6:29:13, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 320/1791 [1:24:40<6:29:13, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 321/1791 [1:24:55<6:28:55, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 321/1791 [1:24:55<6:28:55, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 322/1791 [1:25:11<6:28:40, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 322/1791 [1:25:11<6:28:40, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 323/1791 [1:25:27<6:28:21, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 323/1791 [1:25:27<6:28:21, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 324/1791 [1:25:42<6:28:04, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 324/1791 [1:25:42<6:28:04, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 325/1791 [1:25:58<6:27:47, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 325/1791 [1:25:58<6:27:47, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 326/1791 [1:26:13<6:27:28, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 326/1791 [1:26:13<6:27:28, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 327/1791 [1:26:29<6:27:15, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 327/1791 [1:26:29<6:27:15, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 328/1791 [1:26:46<6:27:01, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 328/1791 [1:26:46<6:27:01, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 329/1791 [1:27:02<6:26:45, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 329/1791 [1:27:02<6:26:45, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 330/1791 [1:27:17<6:26:28, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 330/1791 [1:27:17<6:26:28, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 331/1791 [1:27:33<6:26:11, 0.06it/s, v_num=bc8f] Epoch 0: 18%|█▊ | 331/1791 [1:27:33<6:26:11, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 332/1791 [1:27:48<6:25:53, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 332/1791 [1:27:48<6:25:54, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 333/1791 [1:28:04<6:25:36, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 333/1791 [1:28:04<6:25:36, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 334/1791 [1:28:19<6:25:19, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 334/1791 [1:28:19<6:25:19, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 335/1791 [1:28:35<6:25:00, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▊ | 335/1791 [1:28:35<6:25:00, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 336/1791 [1:28:52<6:24:49, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 336/1791 [1:28:52<6:24:49, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 337/1791 [1:29:07<6:24:33, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 337/1791 [1:29:07<6:24:33, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 338/1791 [1:29:23<6:24:14, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 338/1791 [1:29:23<6:24:14, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 339/1791 [1:29:38<6:23:58, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 339/1791 [1:29:38<6:23:58, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 340/1791 [1:29:54<6:23:42, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 340/1791 [1:29:54<6:23:42, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 341/1791 [1:30:10<6:23:27, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 341/1791 [1:30:10<6:23:27, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 342/1791 [1:30:26<6:23:11, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 342/1791 [1:30:26<6:23:11, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 343/1791 [1:30:42<6:22:53, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 343/1791 [1:30:42<6:22:53, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 344/1791 [1:30:59<6:22:43, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 344/1791 [1:30:59<6:22:43, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 345/1791 [1:31:14<6:22:24, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 345/1791 [1:31:14<6:22:24, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 346/1791 [1:31:30<6:22:08, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 346/1791 [1:31:30<6:22:08, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 347/1791 [1:31:45<6:21:51, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 347/1791 [1:31:45<6:21:51, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 348/1791 [1:32:01<6:21:36, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 348/1791 [1:32:01<6:21:36, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 349/1791 [1:32:17<6:21:21, 0.06it/s, v_num=bc8f] Epoch 0: 19%|█▉ | 349/1791 [1:32:17<6:21:21, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 350/1791 [1:32:34<6:21:08, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 350/1791 [1:32:34<6:21:08, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 351/1791 [1:32:49<6:20:49, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 351/1791 [1:32:49<6:20:49, 0.06it/s, v_num=bc8f]Token indices sequence length is longer than the specified maximum sequence length for this model (1237 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 20%|█▉ | 352/1791 [1:33:06<6:20:39, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 352/1791 [1:33:06<6:20:39, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 353/1791 [1:33:22<6:20:22, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 353/1791 [1:33:22<6:20:22, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 354/1791 [1:33:38<6:20:07, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 354/1791 [1:33:38<6:20:07, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 355/1791 [1:33:54<6:19:51, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 355/1791 [1:33:54<6:19:51, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 356/1791 [1:34:10<6:19:36, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 356/1791 [1:34:10<6:19:36, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 357/1791 [1:34:26<6:19:19, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 357/1791 [1:34:26<6:19:19, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 358/1791 [1:34:41<6:19:03, 0.06it/s, v_num=bc8f] Epoch 0: 20%|█▉ | 358/1791 [1:34:41<6:19:03, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 359/1791 [1:34:56<6:18:44, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 359/1791 [1:34:56<6:18:44, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 360/1791 [1:35:13<6:18:32, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 360/1791 [1:35:13<6:18:32, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 361/1791 [1:35:29<6:18:15, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 361/1791 [1:35:29<6:18:15, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 362/1791 [1:35:45<6:18:00, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 362/1791 [1:35:45<6:18:00, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 363/1791 [1:36:01<6:17:44, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 363/1791 [1:36:01<6:17:44, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 364/1791 [1:36:17<6:17:28, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 364/1791 [1:36:17<6:17:28, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 365/1791 [1:36:32<6:17:09, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 365/1791 [1:36:32<6:17:09, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 366/1791 [1:36:47<6:16:52, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 366/1791 [1:36:47<6:16:52, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 367/1791 [1:37:03<6:16:35, 0.06it/s, v_num=bc8f] Epoch 0: 20%|██ | 367/1791 [1:37:03<6:16:35, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 368/1791 [1:37:20<6:16:22, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 368/1791 [1:37:20<6:16:22, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 369/1791 [1:37:35<6:16:05, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 369/1791 [1:37:35<6:16:05, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 370/1791 [1:37:51<6:15:49, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 370/1791 [1:37:51<6:15:49, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 371/1791 [1:38:07<6:15:34, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 371/1791 [1:38:07<6:15:34, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 372/1791 [1:38:23<6:15:17, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 372/1791 [1:38:23<6:15:17, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 373/1791 [1:38:38<6:15:00, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 373/1791 [1:38:38<6:15:00, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 374/1791 [1:38:54<6:14:43, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 374/1791 [1:38:54<6:14:43, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 375/1791 [1:39:09<6:14:26, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 375/1791 [1:39:09<6:14:26, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 376/1791 [1:39:27<6:14:16, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 376/1791 [1:39:27<6:14:16, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 377/1791 [1:39:43<6:14:00, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 377/1791 [1:39:43<6:14:00, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 378/1791 [1:39:58<6:13:41, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 378/1791 [1:39:58<6:13:41, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 379/1791 [1:40:13<6:13:25, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 379/1791 [1:40:13<6:13:25, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 380/1791 [1:40:29<6:13:09, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██ | 380/1791 [1:40:29<6:13:09, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 381/1791 [1:40:45<6:12:54, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 381/1791 [1:40:45<6:12:54, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 382/1791 [1:41:01<6:12:36, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 382/1791 [1:41:01<6:12:36, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 383/1791 [1:41:16<6:12:19, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 383/1791 [1:41:16<6:12:19, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 384/1791 [1:41:34<6:12:09, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 384/1791 [1:41:34<6:12:09, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 385/1791 [1:41:49<6:11:53, 0.06it/s, v_num=bc8f] Epoch 0: 21%|██▏ | 385/1791 [1:41:49<6:11:53, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 386/1791 [1:42:05<6:11:34, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 386/1791 [1:42:05<6:11:34, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 387/1791 [1:42:20<6:11:17, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 387/1791 [1:42:20<6:11:17, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 388/1791 [1:42:36<6:11:01, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 388/1791 [1:42:36<6:11:01, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 389/1791 [1:42:51<6:10:44, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 389/1791 [1:42:51<6:10:44, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 390/1791 [1:43:07<6:10:27, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 390/1791 [1:43:07<6:10:27, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 391/1791 [1:43:24<6:10:14, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 391/1791 [1:43:24<6:10:14, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 392/1791 [1:43:41<6:10:03, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 392/1791 [1:43:41<6:10:03, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 393/1791 [1:43:57<6:09:47, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 393/1791 [1:43:57<6:09:47, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 394/1791 [1:44:13<6:09:32, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 394/1791 [1:44:13<6:09:32, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 395/1791 [1:44:28<6:09:15, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 395/1791 [1:44:28<6:09:15, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 396/1791 [1:44:43<6:08:56, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 396/1791 [1:44:43<6:08:56, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 397/1791 [1:44:59<6:08:38, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 397/1791 [1:44:59<6:08:38, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 398/1791 [1:45:15<6:08:23, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 398/1791 [1:45:15<6:08:23, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 399/1791 [1:45:30<6:08:06, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 399/1791 [1:45:30<6:08:06, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 400/1791 [1:45:48<6:07:57, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 400/1791 [1:45:48<6:07:57, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 401/1791 [1:46:04<6:07:42, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 401/1791 [1:46:04<6:07:42, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 402/1791 [1:46:20<6:07:27, 0.06it/s, v_num=bc8f] Epoch 0: 22%|██▏ | 402/1791 [1:46:20<6:07:27, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 403/1791 [1:46:36<6:07:10, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 403/1791 [1:46:36<6:07:10, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 404/1791 [1:46:52<6:06:56, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 404/1791 [1:46:52<6:06:56, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 405/1791 [1:47:08<6:06:40, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 405/1791 [1:47:08<6:06:40, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 406/1791 [1:47:24<6:06:25, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 406/1791 [1:47:24<6:06:25, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 407/1791 [1:47:40<6:06:09, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 407/1791 [1:47:40<6:06:09, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 408/1791 [1:47:57<6:05:57, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 408/1791 [1:47:57<6:05:57, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 409/1791 [1:48:12<6:05:39, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 409/1791 [1:48:12<6:05:39, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 410/1791 [1:48:28<6:05:23, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 410/1791 [1:48:28<6:05:23, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 411/1791 [1:48:44<6:05:06, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 411/1791 [1:48:44<6:05:06, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 412/1791 [1:49:00<6:04:51, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 412/1791 [1:49:00<6:04:51, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 413/1791 [1:49:16<6:04:36, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 413/1791 [1:49:16<6:04:36, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 414/1791 [1:49:32<6:04:21, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 414/1791 [1:49:32<6:04:21, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 415/1791 [1:49:49<6:04:06, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 415/1791 [1:49:49<6:04:06, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 416/1791 [1:50:06<6:03:57, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 416/1791 [1:50:06<6:03:57, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 417/1791 [1:50:22<6:03:42, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 417/1791 [1:50:22<6:03:42, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 418/1791 [1:50:38<6:03:24, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 418/1791 [1:50:38<6:03:24, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 419/1791 [1:50:53<6:03:07, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 419/1791 [1:50:53<6:03:07, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 420/1791 [1:51:09<6:02:52, 0.06it/s, v_num=bc8f] Epoch 0: 23%|██▎ | 420/1791 [1:51:09<6:02:52, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 421/1791 [1:51:25<6:02:36, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 421/1791 [1:51:25<6:02:36, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 422/1791 [1:51:41<6:02:18, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 422/1791 [1:51:41<6:02:18, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 423/1791 [1:51:56<6:02:02, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 423/1791 [1:51:56<6:02:02, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 424/1791 [1:52:13<6:01:49, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 424/1791 [1:52:13<6:01:49, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 425/1791 [1:52:28<6:01:31, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▎ | 425/1791 [1:52:28<6:01:31, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 426/1791 [1:52:44<6:01:14, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 426/1791 [1:52:44<6:01:14, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 427/1791 [1:53:00<6:00:59, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 427/1791 [1:53:00<6:00:59, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 428/1791 [1:53:16<6:00:43, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 428/1791 [1:53:16<6:00:43, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 429/1791 [1:53:32<6:00:27, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 429/1791 [1:53:32<6:00:27, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 430/1791 [1:53:47<6:00:11, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 430/1791 [1:53:47<6:00:11, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 431/1791 [1:54:03<5:59:54, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 431/1791 [1:54:03<5:59:54, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 432/1791 [1:54:20<5:59:42, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 432/1791 [1:54:20<5:59:42, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 433/1791 [1:54:36<5:59:25, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 433/1791 [1:54:36<5:59:25, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 434/1791 [1:54:52<5:59:09, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 434/1791 [1:54:52<5:59:09, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 435/1791 [1:55:07<5:58:52, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 435/1791 [1:55:07<5:58:52, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 436/1791 [1:55:23<5:58:36, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 436/1791 [1:55:23<5:58:36, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 437/1791 [1:55:39<5:58:19, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 437/1791 [1:55:39<5:58:19, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 438/1791 [1:55:54<5:58:02, 0.06it/s, v_num=bc8f] Epoch 0: 24%|██▍ | 438/1791 [1:55:54<5:58:02, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 439/1791 [1:56:10<5:57:46, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 439/1791 [1:56:10<5:57:46, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 440/1791 [1:56:27<5:57:35, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 440/1791 [1:56:27<5:57:35, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 441/1791 [1:56:43<5:57:19, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 441/1791 [1:56:43<5:57:19, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 442/1791 [1:56:59<5:57:03, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 442/1791 [1:56:59<5:57:03, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 443/1791 [1:57:14<5:56:46, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 443/1791 [1:57:14<5:56:46, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 444/1791 [1:57:30<5:56:29, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 444/1791 [1:57:30<5:56:29, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 445/1791 [1:57:45<5:56:11, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 445/1791 [1:57:45<5:56:11, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 446/1791 [1:58:01<5:55:54, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 446/1791 [1:58:01<5:55:54, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 447/1791 [1:58:16<5:55:38, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▍ | 447/1791 [1:58:16<5:55:38, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 448/1791 [1:58:34<5:55:26, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 448/1791 [1:58:34<5:55:26, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 449/1791 [1:58:50<5:55:11, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 449/1791 [1:58:50<5:55:11, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 450/1791 [1:59:06<5:54:56, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 450/1791 [1:59:06<5:54:56, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 451/1791 [1:59:21<5:54:39, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 451/1791 [1:59:21<5:54:39, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 452/1791 [1:59:37<5:54:23, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 452/1791 [1:59:37<5:54:23, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 453/1791 [1:59:53<5:54:06, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 453/1791 [1:59:53<5:54:06, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 454/1791 [2:00:08<5:53:49, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 454/1791 [2:00:08<5:53:49, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 455/1791 [2:00:23<5:53:31, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 455/1791 [2:00:23<5:53:31, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 456/1791 [2:00:40<5:53:18, 0.06it/s, v_num=bc8f] Epoch 0: 25%|██▌ | 456/1791 [2:00:40<5:53:18, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 457/1791 [2:00:56<5:53:00, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 457/1791 [2:00:56<5:53:00, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 458/1791 [2:01:11<5:52:43, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 458/1791 [2:01:11<5:52:43, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 459/1791 [2:01:26<5:52:26, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 459/1791 [2:01:26<5:52:26, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 460/1791 [2:01:42<5:52:09, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 460/1791 [2:01:42<5:52:09, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 461/1791 [2:01:57<5:51:51, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 461/1791 [2:01:57<5:51:51, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 462/1791 [2:02:13<5:51:35, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 462/1791 [2:02:13<5:51:35, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 463/1791 [2:02:29<5:51:20, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 463/1791 [2:02:29<5:51:20, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 464/1791 [2:02:46<5:51:08, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 464/1791 [2:02:46<5:51:08, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 465/1791 [2:03:02<5:50:53, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 465/1791 [2:03:02<5:50:53, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 466/1791 [2:03:18<5:50:36, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 466/1791 [2:03:18<5:50:36, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 467/1791 [2:03:34<5:50:19, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 467/1791 [2:03:34<5:50:19, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 468/1791 [2:03:49<5:50:03, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 468/1791 [2:03:49<5:50:03, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 469/1791 [2:04:04<5:49:45, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 469/1791 [2:04:04<5:49:45, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 470/1791 [2:04:20<5:49:29, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▌ | 470/1791 [2:04:20<5:49:29, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 471/1791 [2:04:36<5:49:12, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 471/1791 [2:04:36<5:49:12, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 472/1791 [2:04:53<5:48:59, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 472/1791 [2:04:53<5:48:59, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 473/1791 [2:05:08<5:48:43, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 473/1791 [2:05:08<5:48:43, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 474/1791 [2:05:24<5:48:26, 0.06it/s, v_num=bc8f] Epoch 0: 26%|██▋ | 474/1791 [2:05:24<5:48:26, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 475/1791 [2:05:40<5:48:10, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 475/1791 [2:05:40<5:48:10, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 476/1791 [2:05:55<5:47:53, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 476/1791 [2:05:55<5:47:53, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 477/1791 [2:06:11<5:47:36, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 477/1791 [2:06:11<5:47:36, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 478/1791 [2:06:27<5:47:20, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 478/1791 [2:06:27<5:47:20, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 479/1791 [2:06:42<5:47:04, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 479/1791 [2:06:42<5:47:04, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 480/1791 [2:06:59<5:46:52, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 480/1791 [2:06:59<5:46:52, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 481/1791 [2:07:16<5:46:36, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 481/1791 [2:07:16<5:46:36, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 482/1791 [2:07:31<5:46:21, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 482/1791 [2:07:31<5:46:21, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 483/1791 [2:07:48<5:46:05, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 483/1791 [2:07:48<5:46:05, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 484/1791 [2:08:04<5:45:50, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 484/1791 [2:08:04<5:45:50, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 485/1791 [2:08:19<5:45:33, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 485/1791 [2:08:19<5:45:33, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 486/1791 [2:08:35<5:45:16, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 486/1791 [2:08:35<5:45:16, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 487/1791 [2:08:50<5:45:00, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 487/1791 [2:08:50<5:45:00, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 488/1791 [2:09:08<5:44:49, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 488/1791 [2:09:08<5:44:49, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 489/1791 [2:09:23<5:44:32, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 489/1791 [2:09:23<5:44:32, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 490/1791 [2:09:39<5:44:15, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 490/1791 [2:09:39<5:44:15, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 491/1791 [2:09:55<5:43:59, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 491/1791 [2:09:55<5:43:59, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 492/1791 [2:10:10<5:43:42, 0.06it/s, v_num=bc8f] Epoch 0: 27%|██▋ | 492/1791 [2:10:10<5:43:42, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 493/1791 [2:10:26<5:43:27, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 493/1791 [2:10:26<5:43:27, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 494/1791 [2:10:42<5:43:11, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 494/1791 [2:10:42<5:43:11, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 495/1791 [2:10:57<5:42:53, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 495/1791 [2:10:57<5:42:53, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 496/1791 [2:11:15<5:42:42, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 496/1791 [2:11:15<5:42:42, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 497/1791 [2:11:31<5:42:26, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 497/1791 [2:11:31<5:42:26, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 498/1791 [2:11:47<5:42:09, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 498/1791 [2:11:47<5:42:09, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 499/1791 [2:12:02<5:41:52, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 499/1791 [2:12:02<5:41:52, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 500/1791 [2:12:17<5:41:35, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 500/1791 [2:12:17<5:41:35, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 501/1791 [2:12:33<5:41:17, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 501/1791 [2:12:33<5:41:17, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 502/1791 [2:12:48<5:41:01, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 502/1791 [2:12:48<5:41:01, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 503/1791 [2:13:04<5:40:44, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 503/1791 [2:13:04<5:40:44, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 504/1791 [2:13:21<5:40:32, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 504/1791 [2:13:21<5:40:32, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 505/1791 [2:13:37<5:40:15, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 505/1791 [2:13:37<5:40:15, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 506/1791 [2:13:52<5:39:59, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 506/1791 [2:13:52<5:39:59, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 507/1791 [2:14:08<5:39:43, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 507/1791 [2:14:08<5:39:43, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 508/1791 [2:14:24<5:39:26, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 508/1791 [2:14:24<5:39:26, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 509/1791 [2:14:39<5:39:09, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 509/1791 [2:14:39<5:39:09, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 510/1791 [2:14:54<5:38:51, 0.06it/s, v_num=bc8f] Epoch 0: 28%|██▊ | 510/1791 [2:14:54<5:38:51, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 511/1791 [2:15:10<5:38:35, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 511/1791 [2:15:10<5:38:35, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 512/1791 [2:15:27<5:38:22, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 512/1791 [2:15:27<5:38:22, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 513/1791 [2:15:43<5:38:07, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 513/1791 [2:15:43<5:38:07, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 514/1791 [2:15:59<5:37:51, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▊ | 514/1791 [2:15:59<5:37:51, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 515/1791 [2:16:15<5:37:36, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 515/1791 [2:16:15<5:37:36, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 516/1791 [2:16:31<5:37:19, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 516/1791 [2:16:31<5:37:19, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 517/1791 [2:16:46<5:37:02, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 517/1791 [2:16:46<5:37:02, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 518/1791 [2:17:01<5:36:45, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 518/1791 [2:17:01<5:36:45, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 519/1791 [2:17:17<5:36:28, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 519/1791 [2:17:17<5:36:28, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 520/1791 [2:17:34<5:36:16, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 520/1791 [2:17:34<5:36:16, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 521/1791 [2:17:51<5:36:02, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 521/1791 [2:17:51<5:36:02, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 522/1791 [2:18:06<5:35:45, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 522/1791 [2:18:06<5:35:45, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 523/1791 [2:18:23<5:35:30, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 523/1791 [2:18:23<5:35:30, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 524/1791 [2:18:39<5:35:15, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 524/1791 [2:18:39<5:35:15, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 525/1791 [2:18:55<5:35:00, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 525/1791 [2:18:55<5:35:00, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 526/1791 [2:19:12<5:34:46, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 526/1791 [2:19:12<5:34:46, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 527/1791 [2:19:28<5:34:30, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 527/1791 [2:19:28<5:34:30, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 528/1791 [2:19:46<5:34:19, 0.06it/s, v_num=bc8f] Epoch 0: 29%|██▉ | 528/1791 [2:19:46<5:34:19, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 529/1791 [2:20:02<5:34:04, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 529/1791 [2:20:02<5:34:04, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 530/1791 [2:20:18<5:33:48, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 530/1791 [2:20:18<5:33:48, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 531/1791 [2:20:34<5:33:33, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 531/1791 [2:20:34<5:33:33, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 532/1791 [2:20:49<5:33:15, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 532/1791 [2:20:49<5:33:15, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 533/1791 [2:21:05<5:32:59, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 533/1791 [2:21:05<5:32:59, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 534/1791 [2:21:20<5:32:43, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 534/1791 [2:21:20<5:32:43, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 535/1791 [2:21:36<5:32:27, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 535/1791 [2:21:36<5:32:27, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 536/1791 [2:21:53<5:32:14, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 536/1791 [2:21:53<5:32:14, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 537/1791 [2:22:09<5:31:58, 0.06it/s, v_num=bc8f] Epoch 0: 30%|██▉ | 537/1791 [2:22:09<5:31:58, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 538/1791 [2:22:25<5:31:42, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 538/1791 [2:22:25<5:31:42, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 539/1791 [2:22:41<5:31:27, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 539/1791 [2:22:41<5:31:27, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 540/1791 [2:22:57<5:31:11, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 540/1791 [2:22:57<5:31:11, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 541/1791 [2:23:13<5:30:54, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 541/1791 [2:23:13<5:30:54, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 542/1791 [2:23:29<5:30:39, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 542/1791 [2:23:29<5:30:39, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 543/1791 [2:23:44<5:30:22, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 543/1791 [2:23:44<5:30:22, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 544/1791 [2:24:02<5:30:10, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 544/1791 [2:24:02<5:30:10, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 545/1791 [2:24:17<5:29:53, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 545/1791 [2:24:17<5:29:53, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 546/1791 [2:24:33<5:29:38, 0.06it/s, v_num=bc8f] Epoch 0: 30%|███ | 546/1791 [2:24:33<5:29:38, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 547/1791 [2:24:49<5:29:21, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 547/1791 [2:24:49<5:29:21, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 548/1791 [2:25:05<5:29:06, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 548/1791 [2:25:05<5:29:06, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 549/1791 [2:25:21<5:28:49, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 549/1791 [2:25:21<5:28:49, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 550/1791 [2:25:36<5:28:32, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 550/1791 [2:25:36<5:28:32, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 551/1791 [2:25:52<5:28:17, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 551/1791 [2:25:52<5:28:17, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 552/1791 [2:26:09<5:28:04, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 552/1791 [2:26:09<5:28:04, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 553/1791 [2:26:25<5:27:48, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 553/1791 [2:26:25<5:27:48, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 554/1791 [2:26:41<5:27:32, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 554/1791 [2:26:41<5:27:32, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 555/1791 [2:26:58<5:27:18, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 555/1791 [2:26:58<5:27:18, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 556/1791 [2:27:14<5:27:03, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 556/1791 [2:27:14<5:27:03, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 557/1791 [2:27:30<5:26:48, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 557/1791 [2:27:30<5:26:48, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 558/1791 [2:27:47<5:26:33, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 558/1791 [2:27:47<5:26:33, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 559/1791 [2:28:03<5:26:18, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███ | 559/1791 [2:28:03<5:26:18, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 560/1791 [2:28:21<5:26:07, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 560/1791 [2:28:21<5:26:07, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 561/1791 [2:28:37<5:25:51, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 561/1791 [2:28:37<5:25:51, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 562/1791 [2:28:53<5:25:35, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 562/1791 [2:28:53<5:25:35, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 563/1791 [2:29:09<5:25:20, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 563/1791 [2:29:09<5:25:20, 0.06it/s, v_num=bc8f]Token indices sequence length is longer than the specified maximum sequence length for this model (1439 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 31%|███▏ | 564/1791 [2:29:26<5:25:06, 0.06it/s, v_num=bc8f] Epoch 0: 31%|███▏ | 564/1791 [2:29:26<5:25:06, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 565/1791 [2:29:42<5:24:50, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 565/1791 [2:29:42<5:24:50, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 566/1791 [2:29:58<5:24:34, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 566/1791 [2:29:58<5:24:34, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 567/1791 [2:30:13<5:24:18, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 567/1791 [2:30:13<5:24:18, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 568/1791 [2:30:31<5:24:05, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 568/1791 [2:30:31<5:24:05, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 569/1791 [2:30:46<5:23:49, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 569/1791 [2:30:46<5:23:49, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 570/1791 [2:31:02<5:23:32, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 570/1791 [2:31:02<5:23:32, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 571/1791 [2:31:18<5:23:16, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 571/1791 [2:31:18<5:23:16, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 572/1791 [2:31:33<5:22:59, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 572/1791 [2:31:33<5:22:59, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 573/1791 [2:31:48<5:22:42, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 573/1791 [2:31:48<5:22:42, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 574/1791 [2:32:04<5:22:25, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 574/1791 [2:32:04<5:22:25, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 575/1791 [2:32:20<5:22:10, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 575/1791 [2:32:20<5:22:10, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 576/1791 [2:32:37<5:21:56, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 576/1791 [2:32:37<5:21:56, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 577/1791 [2:32:53<5:21:40, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 577/1791 [2:32:53<5:21:40, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 578/1791 [2:33:08<5:21:23, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 578/1791 [2:33:08<5:21:23, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 579/1791 [2:33:24<5:21:08, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 579/1791 [2:33:24<5:21:08, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 580/1791 [2:33:40<5:20:51, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 580/1791 [2:33:40<5:20:51, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 581/1791 [2:33:56<5:20:35, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 581/1791 [2:33:56<5:20:35, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 582/1791 [2:34:12<5:20:19, 0.06it/s, v_num=bc8f] Epoch 0: 32%|███▏ | 582/1791 [2:34:12<5:20:19, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 583/1791 [2:34:27<5:20:02, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 583/1791 [2:34:27<5:20:02, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 584/1791 [2:34:44<5:19:48, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 584/1791 [2:34:44<5:19:48, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 585/1791 [2:35:00<5:19:34, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 585/1791 [2:35:00<5:19:34, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 586/1791 [2:35:16<5:19:17, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 586/1791 [2:35:16<5:19:17, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 587/1791 [2:35:31<5:19:00, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 587/1791 [2:35:31<5:19:00, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 588/1791 [2:35:46<5:18:42, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 588/1791 [2:35:46<5:18:42, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 589/1791 [2:36:02<5:18:26, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 589/1791 [2:36:02<5:18:26, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 590/1791 [2:36:18<5:18:11, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 590/1791 [2:36:18<5:18:11, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 591/1791 [2:36:34<5:17:55, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 591/1791 [2:36:34<5:17:55, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 592/1791 [2:36:52<5:17:43, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 592/1791 [2:36:52<5:17:43, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 593/1791 [2:37:08<5:17:27, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 593/1791 [2:37:08<5:17:27, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 594/1791 [2:37:24<5:17:12, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 594/1791 [2:37:24<5:17:12, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 595/1791 [2:37:40<5:16:57, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 595/1791 [2:37:40<5:16:57, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 596/1791 [2:37:57<5:16:41, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 596/1791 [2:37:57<5:16:41, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 597/1791 [2:38:13<5:16:26, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 597/1791 [2:38:13<5:16:26, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 598/1791 [2:38:29<5:16:10, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 598/1791 [2:38:29<5:16:10, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 599/1791 [2:38:44<5:15:53, 0.06it/s, v_num=bc8f] Epoch 0: 33%|███▎ | 599/1791 [2:38:44<5:15:53, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 600/1791 [2:39:02<5:15:41, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 600/1791 [2:39:02<5:15:41, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 601/1791 [2:39:17<5:15:24, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 601/1791 [2:39:17<5:15:25, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 602/1791 [2:39:33<5:15:09, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 602/1791 [2:39:33<5:15:09, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 603/1791 [2:39:49<5:14:52, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 603/1791 [2:39:49<5:14:52, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 604/1791 [2:40:05<5:14:37, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▎ | 604/1791 [2:40:05<5:14:37, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 605/1791 [2:40:21<5:14:21, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 605/1791 [2:40:21<5:14:21, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 606/1791 [2:40:36<5:14:04, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 606/1791 [2:40:36<5:14:04, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 607/1791 [2:40:52<5:13:47, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 607/1791 [2:40:52<5:13:47, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 608/1791 [2:41:09<5:13:33, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 608/1791 [2:41:09<5:13:33, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 609/1791 [2:41:25<5:13:18, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 609/1791 [2:41:25<5:13:18, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 610/1791 [2:41:41<5:13:01, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 610/1791 [2:41:41<5:13:01, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 611/1791 [2:41:56<5:12:44, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 611/1791 [2:41:56<5:12:44, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 612/1791 [2:42:12<5:12:30, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 612/1791 [2:42:12<5:12:30, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 613/1791 [2:42:29<5:12:15, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 613/1791 [2:42:29<5:12:15, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 614/1791 [2:42:45<5:11:59, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 614/1791 [2:42:45<5:11:59, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 615/1791 [2:43:01<5:11:43, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 615/1791 [2:43:01<5:11:43, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 616/1791 [2:43:18<5:11:30, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 616/1791 [2:43:18<5:11:30, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 617/1791 [2:43:34<5:11:14, 0.06it/s, v_num=bc8f] Epoch 0: 34%|███▍ | 617/1791 [2:43:34<5:11:14, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 618/1791 [2:43:50<5:10:58, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 618/1791 [2:43:50<5:10:58, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 619/1791 [2:44:05<5:10:41, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 619/1791 [2:44:05<5:10:41, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 620/1791 [2:44:21<5:10:24, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 620/1791 [2:44:21<5:10:24, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 621/1791 [2:44:36<5:10:08, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 621/1791 [2:44:36<5:10:08, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 622/1791 [2:44:53<5:09:53, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 622/1791 [2:44:53<5:09:53, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 623/1791 [2:45:09<5:09:37, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 623/1791 [2:45:09<5:09:37, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 624/1791 [2:45:26<5:09:23, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 624/1791 [2:45:26<5:09:23, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 625/1791 [2:45:42<5:09:07, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 625/1791 [2:45:42<5:09:07, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 626/1791 [2:45:57<5:08:51, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▍ | 626/1791 [2:45:57<5:08:51, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 627/1791 [2:46:13<5:08:35, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 627/1791 [2:46:13<5:08:35, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 628/1791 [2:46:29<5:08:19, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 628/1791 [2:46:29<5:08:19, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 629/1791 [2:46:44<5:08:02, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 629/1791 [2:46:44<5:08:02, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 630/1791 [2:47:00<5:07:47, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 630/1791 [2:47:00<5:07:47, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 631/1791 [2:47:16<5:07:30, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 631/1791 [2:47:16<5:07:30, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 632/1791 [2:47:34<5:07:17, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 632/1791 [2:47:34<5:07:17, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 633/1791 [2:47:49<5:07:01, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 633/1791 [2:47:49<5:07:01, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 634/1791 [2:48:05<5:06:45, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 634/1791 [2:48:05<5:06:45, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 635/1791 [2:48:21<5:06:29, 0.06it/s, v_num=bc8f] Epoch 0: 35%|███▌ | 635/1791 [2:48:21<5:06:29, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 636/1791 [2:48:37<5:06:14, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 636/1791 [2:48:37<5:06:14, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 637/1791 [2:48:53<5:05:58, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 637/1791 [2:48:53<5:05:58, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 638/1791 [2:49:09<5:05:42, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 638/1791 [2:49:09<5:05:42, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 639/1791 [2:49:25<5:05:26, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 639/1791 [2:49:25<5:05:26, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 640/1791 [2:49:42<5:05:13, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 640/1791 [2:49:42<5:05:13, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 641/1791 [2:49:58<5:04:56, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 641/1791 [2:49:58<5:04:56, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 642/1791 [2:50:13<5:04:39, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 642/1791 [2:50:13<5:04:39, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 643/1791 [2:50:29<5:04:22, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 643/1791 [2:50:29<5:04:22, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 644/1791 [2:50:44<5:04:06, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 644/1791 [2:50:44<5:04:06, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 645/1791 [2:51:00<5:03:50, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 645/1791 [2:51:00<5:03:50, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 646/1791 [2:51:16<5:03:33, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 646/1791 [2:51:16<5:03:33, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 647/1791 [2:51:31<5:03:17, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 647/1791 [2:51:31<5:03:17, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 648/1791 [2:51:49<5:03:05, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 648/1791 [2:51:49<5:03:05, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 649/1791 [2:52:05<5:02:48, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▌ | 649/1791 [2:52:05<5:02:48, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 650/1791 [2:52:21<5:02:33, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 650/1791 [2:52:21<5:02:33, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 651/1791 [2:52:37<5:02:17, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 651/1791 [2:52:37<5:02:17, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 652/1791 [2:52:53<5:02:02, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 652/1791 [2:52:53<5:02:02, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 653/1791 [2:53:09<5:01:46, 0.06it/s, v_num=bc8f] Epoch 0: 36%|███▋ | 653/1791 [2:53:09<5:01:46, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 654/1791 [2:53:25<5:01:30, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 654/1791 [2:53:25<5:01:30, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 655/1791 [2:53:41<5:01:14, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 655/1791 [2:53:41<5:01:14, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 656/1791 [2:53:58<5:01:00, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 656/1791 [2:53:58<5:01:00, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 657/1791 [2:54:14<5:00:44, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 657/1791 [2:54:14<5:00:44, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 658/1791 [2:54:30<5:00:28, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 658/1791 [2:54:30<5:00:28, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 659/1791 [2:54:45<5:00:12, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 659/1791 [2:54:45<5:00:12, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 660/1791 [2:55:01<4:59:55, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 660/1791 [2:55:01<4:59:55, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 661/1791 [2:55:17<4:59:39, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 661/1791 [2:55:17<4:59:39, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 662/1791 [2:55:32<4:59:23, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 662/1791 [2:55:32<4:59:23, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 663/1791 [2:55:49<4:59:08, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 663/1791 [2:55:49<4:59:08, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 664/1791 [2:56:05<4:58:53, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 664/1791 [2:56:05<4:58:53, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 665/1791 [2:56:21<4:58:36, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 665/1791 [2:56:21<4:58:36, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 666/1791 [2:56:36<4:58:20, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 666/1791 [2:56:36<4:58:20, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 667/1791 [2:56:53<4:58:05, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 667/1791 [2:56:53<4:58:05, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 668/1791 [2:57:08<4:57:48, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 668/1791 [2:57:08<4:57:48, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 669/1791 [2:57:24<4:57:31, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 669/1791 [2:57:24<4:57:31, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 670/1791 [2:57:39<4:57:15, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 670/1791 [2:57:39<4:57:15, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 671/1791 [2:57:55<4:56:58, 0.06it/s, v_num=bc8f] Epoch 0: 37%|███▋ | 671/1791 [2:57:55<4:56:58, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 672/1791 [2:58:12<4:56:45, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 672/1791 [2:58:12<4:56:45, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 673/1791 [2:58:29<4:56:30, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 673/1791 [2:58:29<4:56:30, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 674/1791 [2:58:45<4:56:14, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 674/1791 [2:58:45<4:56:14, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 675/1791 [2:59:00<4:55:58, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 675/1791 [2:59:00<4:55:58, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 676/1791 [2:59:16<4:55:41, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 676/1791 [2:59:16<4:55:41, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 677/1791 [2:59:32<4:55:25, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 677/1791 [2:59:32<4:55:25, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 678/1791 [2:59:48<4:55:09, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 678/1791 [2:59:48<4:55:09, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 679/1791 [3:00:04<4:54:54, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 679/1791 [3:00:04<4:54:54, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 680/1791 [3:00:21<4:54:40, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 680/1791 [3:00:21<4:54:40, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 681/1791 [3:00:37<4:54:24, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 681/1791 [3:00:37<4:54:24, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 682/1791 [3:00:53<4:54:08, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 682/1791 [3:00:53<4:54:08, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 683/1791 [3:01:09<4:53:52, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 683/1791 [3:01:09<4:53:52, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 684/1791 [3:01:24<4:53:36, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 684/1791 [3:01:24<4:53:36, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 685/1791 [3:01:40<4:53:19, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 685/1791 [3:01:40<4:53:19, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 686/1791 [3:01:56<4:53:03, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 686/1791 [3:01:56<4:53:03, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 687/1791 [3:02:11<4:52:47, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 687/1791 [3:02:11<4:52:47, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 688/1791 [3:02:29<4:52:33, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 688/1791 [3:02:29<4:52:33, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 689/1791 [3:02:45<4:52:17, 0.06it/s, v_num=bc8f] Epoch 0: 38%|███▊ | 689/1791 [3:02:45<4:52:17, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 690/1791 [3:03:01<4:52:02, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 690/1791 [3:03:01<4:52:02, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 691/1791 [3:03:17<4:51:47, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 691/1791 [3:03:17<4:51:47, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 692/1791 [3:03:32<4:51:29, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 692/1791 [3:03:32<4:51:29, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 693/1791 [3:03:48<4:51:12, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 693/1791 [3:03:48<4:51:12, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 694/1791 [3:04:03<4:50:56, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▊ | 694/1791 [3:04:03<4:50:56, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 695/1791 [3:04:19<4:50:40, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 695/1791 [3:04:19<4:50:40, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 696/1791 [3:04:37<4:50:27, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 696/1791 [3:04:37<4:50:27, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 697/1791 [3:04:52<4:50:11, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 697/1791 [3:04:52<4:50:11, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 698/1791 [3:05:08<4:49:55, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 698/1791 [3:05:08<4:49:55, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 699/1791 [3:05:24<4:49:38, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 699/1791 [3:05:24<4:49:38, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 700/1791 [3:05:39<4:49:21, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 700/1791 [3:05:39<4:49:21, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 701/1791 [3:05:55<4:49:05, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 701/1791 [3:05:55<4:49:05, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 702/1791 [3:06:11<4:48:50, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 702/1791 [3:06:11<4:48:50, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 703/1791 [3:06:27<4:48:34, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 703/1791 [3:06:27<4:48:34, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 704/1791 [3:06:45<4:48:21, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 704/1791 [3:06:45<4:48:21, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 705/1791 [3:07:01<4:48:05, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 705/1791 [3:07:01<4:48:05, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 706/1791 [3:07:16<4:47:49, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 706/1791 [3:07:16<4:47:49, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 707/1791 [3:07:32<4:47:33, 0.06it/s, v_num=bc8f] Epoch 0: 39%|███▉ | 707/1791 [3:07:32<4:47:33, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 708/1791 [3:07:48<4:47:16, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 708/1791 [3:07:48<4:47:16, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 709/1791 [3:08:04<4:47:01, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 709/1791 [3:08:04<4:47:01, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 710/1791 [3:08:20<4:46:44, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 710/1791 [3:08:20<4:46:44, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 711/1791 [3:08:36<4:46:29, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 711/1791 [3:08:36<4:46:29, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 712/1791 [3:08:53<4:46:14, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 712/1791 [3:08:53<4:46:14, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 713/1791 [3:09:08<4:45:58, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 713/1791 [3:09:08<4:45:58, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 714/1791 [3:09:24<4:45:42, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 714/1791 [3:09:24<4:45:42, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 715/1791 [3:09:40<4:45:26, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 715/1791 [3:09:40<4:45:26, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 716/1791 [3:09:56<4:45:10, 0.06it/s, v_num=bc8f] Epoch 0: 40%|███▉ | 716/1791 [3:09:56<4:45:10, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 717/1791 [3:10:12<4:44:54, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 717/1791 [3:10:12<4:44:54, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 718/1791 [3:10:28<4:44:38, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 718/1791 [3:10:28<4:44:38, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 719/1791 [3:10:44<4:44:23, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 719/1791 [3:10:44<4:44:23, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 720/1791 [3:11:01<4:44:09, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 720/1791 [3:11:01<4:44:09, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 721/1791 [3:11:17<4:43:53, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 721/1791 [3:11:17<4:43:53, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 722/1791 [3:11:33<4:43:37, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 722/1791 [3:11:33<4:43:37, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 723/1791 [3:11:48<4:43:20, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 723/1791 [3:11:48<4:43:20, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 724/1791 [3:12:05<4:43:05, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 724/1791 [3:12:05<4:43:05, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 725/1791 [3:12:21<4:42:50, 0.06it/s, v_num=bc8f] Epoch 0: 40%|████ | 725/1791 [3:12:21<4:42:50, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 726/1791 [3:12:37<4:42:34, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 726/1791 [3:12:37<4:42:34, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 727/1791 [3:12:53<4:42:18, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 727/1791 [3:12:53<4:42:18, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 728/1791 [3:13:11<4:42:05, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 728/1791 [3:13:11<4:42:05, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 729/1791 [3:13:27<4:41:49, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 729/1791 [3:13:27<4:41:49, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 730/1791 [3:13:43<4:41:33, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 730/1791 [3:13:43<4:41:33, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 731/1791 [3:13:59<4:41:17, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 731/1791 [3:13:59<4:41:17, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 732/1791 [3:14:15<4:41:01, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 732/1791 [3:14:15<4:41:01, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 733/1791 [3:14:31<4:40:46, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 733/1791 [3:14:31<4:40:46, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 734/1791 [3:14:47<4:40:30, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 734/1791 [3:14:47<4:40:30, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 735/1791 [3:15:02<4:40:13, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 735/1791 [3:15:02<4:40:13, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 736/1791 [3:15:20<4:40:00, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 736/1791 [3:15:20<4:40:00, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 737/1791 [3:15:36<4:39:44, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 737/1791 [3:15:36<4:39:44, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 738/1791 [3:15:53<4:39:29, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████ | 738/1791 [3:15:53<4:39:29, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 739/1791 [3:16:08<4:39:13, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 739/1791 [3:16:08<4:39:13, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 740/1791 [3:16:24<4:38:57, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 740/1791 [3:16:24<4:38:57, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 741/1791 [3:16:40<4:38:41, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 741/1791 [3:16:40<4:38:41, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 742/1791 [3:16:56<4:38:25, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 742/1791 [3:16:56<4:38:25, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 743/1791 [3:17:12<4:38:09, 0.06it/s, v_num=bc8f] Epoch 0: 41%|████▏ | 743/1791 [3:17:12<4:38:09, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 744/1791 [3:17:29<4:37:55, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 744/1791 [3:17:29<4:37:55, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 745/1791 [3:17:44<4:37:37, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 745/1791 [3:17:44<4:37:37, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 746/1791 [3:18:00<4:37:22, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 746/1791 [3:18:00<4:37:22, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 747/1791 [3:18:16<4:37:06, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 747/1791 [3:18:16<4:37:06, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 748/1791 [3:18:31<4:36:49, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 748/1791 [3:18:31<4:36:49, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 749/1791 [3:18:47<4:36:33, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 749/1791 [3:18:47<4:36:33, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 750/1791 [3:19:02<4:36:16, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 750/1791 [3:19:02<4:36:16, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 751/1791 [3:19:18<4:36:00, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 751/1791 [3:19:18<4:36:00, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 752/1791 [3:19:36<4:35:46, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 752/1791 [3:19:36<4:35:46, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 753/1791 [3:19:51<4:35:30, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 753/1791 [3:19:51<4:35:30, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 754/1791 [3:20:08<4:35:15, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 754/1791 [3:20:08<4:35:15, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 755/1791 [3:20:24<4:34:59, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 755/1791 [3:20:24<4:34:59, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 756/1791 [3:20:40<4:34:43, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 756/1791 [3:20:40<4:34:43, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 757/1791 [3:20:56<4:34:28, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 757/1791 [3:20:56<4:34:28, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 758/1791 [3:21:13<4:34:13, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 758/1791 [3:21:13<4:34:13, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 759/1791 [3:21:29<4:33:57, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 759/1791 [3:21:29<4:33:57, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 760/1791 [3:21:47<4:33:44, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 760/1791 [3:21:47<4:33:44, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 761/1791 [3:22:03<4:33:28, 0.06it/s, v_num=bc8f] Epoch 0: 42%|████▏ | 761/1791 [3:22:03<4:33:28, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 762/1791 [3:22:19<4:33:13, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 762/1791 [3:22:19<4:33:13, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 763/1791 [3:22:35<4:32:57, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 763/1791 [3:22:35<4:32:57, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 764/1791 [3:22:51<4:32:41, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 764/1791 [3:22:51<4:32:41, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 765/1791 [3:23:07<4:32:25, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 765/1791 [3:23:07<4:32:25, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 766/1791 [3:23:23<4:32:09, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 766/1791 [3:23:23<4:32:09, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 767/1791 [3:23:38<4:31:52, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 767/1791 [3:23:38<4:31:52, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 768/1791 [3:23:55<4:31:38, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 768/1791 [3:23:55<4:31:38, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 769/1791 [3:24:11<4:31:21, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 769/1791 [3:24:11<4:31:21, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 770/1791 [3:24:27<4:31:05, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 770/1791 [3:24:27<4:31:05, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 771/1791 [3:24:42<4:30:49, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 771/1791 [3:24:42<4:30:49, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 772/1791 [3:24:58<4:30:33, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 772/1791 [3:24:58<4:30:33, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 773/1791 [3:25:14<4:30:17, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 773/1791 [3:25:14<4:30:17, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 774/1791 [3:25:30<4:30:01, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 774/1791 [3:25:30<4:30:01, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 775/1791 [3:25:46<4:29:45, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 775/1791 [3:25:46<4:29:45, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 776/1791 [3:26:03<4:29:31, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 776/1791 [3:26:03<4:29:31, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 777/1791 [3:26:19<4:29:15, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 777/1791 [3:26:19<4:29:15, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 778/1791 [3:26:36<4:29:00, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 778/1791 [3:26:36<4:29:00, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 779/1791 [3:26:52<4:28:45, 0.06it/s, v_num=bc8f] Epoch 0: 43%|████▎ | 779/1791 [3:26:52<4:28:45, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 780/1791 [3:27:07<4:28:28, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 780/1791 [3:27:07<4:28:28, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 781/1791 [3:27:23<4:28:12, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 781/1791 [3:27:23<4:28:12, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 782/1791 [3:27:39<4:27:56, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 782/1791 [3:27:39<4:27:56, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 783/1791 [3:27:55<4:27:40, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▎ | 783/1791 [3:27:55<4:27:40, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 784/1791 [3:28:12<4:27:25, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 784/1791 [3:28:12<4:27:25, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 785/1791 [3:28:27<4:27:08, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 785/1791 [3:28:27<4:27:08, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 786/1791 [3:28:43<4:26:52, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 786/1791 [3:28:43<4:26:52, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 787/1791 [3:28:59<4:26:36, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 787/1791 [3:28:59<4:26:36, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 788/1791 [3:29:15<4:26:21, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 788/1791 [3:29:15<4:26:21, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 789/1791 [3:29:30<4:26:04, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 789/1791 [3:29:30<4:26:04, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 790/1791 [3:29:47<4:25:49, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 790/1791 [3:29:47<4:25:49, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 791/1791 [3:30:03<4:25:33, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 791/1791 [3:30:03<4:25:33, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 792/1791 [3:30:20<4:25:18, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 792/1791 [3:30:20<4:25:18, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 793/1791 [3:30:36<4:25:02, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 793/1791 [3:30:36<4:25:02, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 794/1791 [3:30:51<4:24:46, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 794/1791 [3:30:51<4:24:46, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 795/1791 [3:31:07<4:24:30, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 795/1791 [3:31:07<4:24:30, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 796/1791 [3:31:23<4:24:14, 0.06it/s, v_num=bc8f] Epoch 0: 44%|████▍ | 796/1791 [3:31:23<4:24:14, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 797/1791 [3:31:39<4:23:58, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 797/1791 [3:31:39<4:23:58, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 798/1791 [3:31:55<4:23:42, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 798/1791 [3:31:55<4:23:42, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 799/1791 [3:32:11<4:23:26, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 799/1791 [3:32:11<4:23:26, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 800/1791 [3:32:29<4:23:13, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 800/1791 [3:32:29<4:23:13, 0.06it/s, v_num=bc8f]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. +✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 + Epoch 0: 45%|████▍ | 801/1791 [3:32:53<4:23:07, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 801/1791 [3:32:53<4:23:07, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 802/1791 [3:33:09<4:22:51, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 802/1791 [3:33:09<4:22:51, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 803/1791 [3:33:25<4:22:36, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 803/1791 [3:33:25<4:22:36, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 804/1791 [3:33:42<4:22:20, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 804/1791 [3:33:42<4:22:20, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 805/1791 [3:33:57<4:22:04, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▍ | 805/1791 [3:33:57<4:22:04, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 806/1791 [3:34:14<4:21:49, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 806/1791 [3:34:14<4:21:49, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 807/1791 [3:34:30<4:21:33, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 807/1791 [3:34:30<4:21:33, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 808/1791 [3:34:47<4:21:18, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 808/1791 [3:34:47<4:21:18, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 809/1791 [3:35:03<4:21:02, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 809/1791 [3:35:03<4:21:02, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 810/1791 [3:35:19<4:20:46, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 810/1791 [3:35:19<4:20:46, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 811/1791 [3:35:34<4:20:30, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 811/1791 [3:35:34<4:20:30, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 812/1791 [3:35:50<4:20:13, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 812/1791 [3:35:50<4:20:13, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 813/1791 [3:36:05<4:19:57, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 813/1791 [3:36:05<4:19:57, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 814/1791 [3:36:21<4:19:40, 0.06it/s, v_num=bc8f] Epoch 0: 45%|████▌ | 814/1791 [3:36:21<4:19:40, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 815/1791 [3:36:36<4:19:23, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 815/1791 [3:36:36<4:19:23, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 816/1791 [3:36:53<4:19:09, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 816/1791 [3:36:53<4:19:09, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 817/1791 [3:37:08<4:18:52, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 817/1791 [3:37:08<4:18:52, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 818/1791 [3:37:23<4:18:35, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 818/1791 [3:37:23<4:18:35, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 819/1791 [3:37:39<4:18:19, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 819/1791 [3:37:39<4:18:19, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 820/1791 [3:37:54<4:18:02, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 820/1791 [3:37:54<4:18:02, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 821/1791 [3:38:10<4:17:46, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 821/1791 [3:38:10<4:17:46, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 822/1791 [3:38:26<4:17:29, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 822/1791 [3:38:26<4:17:29, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 823/1791 [3:38:41<4:17:13, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 823/1791 [3:38:41<4:17:13, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 824/1791 [3:38:59<4:16:59, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 824/1791 [3:38:59<4:16:59, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 825/1791 [3:39:14<4:16:42, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 825/1791 [3:39:14<4:16:42, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 826/1791 [3:39:29<4:16:26, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 826/1791 [3:39:29<4:16:26, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 827/1791 [3:39:45<4:16:09, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 827/1791 [3:39:45<4:16:09, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 828/1791 [3:40:01<4:15:53, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▌ | 828/1791 [3:40:01<4:15:53, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 829/1791 [3:40:16<4:15:37, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 829/1791 [3:40:16<4:15:37, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 830/1791 [3:40:32<4:15:20, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 830/1791 [3:40:32<4:15:20, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 831/1791 [3:40:47<4:15:04, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 831/1791 [3:40:47<4:15:04, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 832/1791 [3:41:04<4:14:49, 0.06it/s, v_num=bc8f] Epoch 0: 46%|████▋ | 832/1791 [3:41:04<4:14:49, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 833/1791 [3:41:19<4:14:32, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 833/1791 [3:41:19<4:14:32, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 834/1791 [3:41:36<4:14:17, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 834/1791 [3:41:36<4:14:17, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 835/1791 [3:41:52<4:14:01, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 835/1791 [3:41:52<4:14:01, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 836/1791 [3:42:07<4:13:45, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 836/1791 [3:42:07<4:13:45, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 837/1791 [3:42:23<4:13:28, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 837/1791 [3:42:23<4:13:28, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 838/1791 [3:42:39<4:13:12, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 838/1791 [3:42:39<4:13:12, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 839/1791 [3:42:55<4:12:56, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 839/1791 [3:42:55<4:12:56, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 840/1791 [3:43:11<4:12:41, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 840/1791 [3:43:11<4:12:41, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 841/1791 [3:43:27<4:12:25, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 841/1791 [3:43:27<4:12:25, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 842/1791 [3:43:43<4:12:09, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 842/1791 [3:43:43<4:12:09, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 843/1791 [3:43:59<4:11:53, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 843/1791 [3:43:59<4:11:53, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 844/1791 [3:44:15<4:11:37, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 844/1791 [3:44:15<4:11:37, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 845/1791 [3:44:30<4:11:20, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 845/1791 [3:44:30<4:11:20, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 846/1791 [3:44:46<4:11:04, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 846/1791 [3:44:46<4:11:04, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 847/1791 [3:45:01<4:10:48, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 847/1791 [3:45:01<4:10:48, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 848/1791 [3:45:19<4:10:33, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 848/1791 [3:45:19<4:10:33, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 849/1791 [3:45:34<4:10:17, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 849/1791 [3:45:34<4:10:17, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 850/1791 [3:45:50<4:10:01, 0.06it/s, v_num=bc8f] Epoch 0: 47%|████▋ | 850/1791 [3:45:50<4:10:01, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 851/1791 [3:46:06<4:09:45, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 851/1791 [3:46:06<4:09:45, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 852/1791 [3:46:22<4:09:29, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 852/1791 [3:46:22<4:09:29, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 853/1791 [3:46:38<4:09:13, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 853/1791 [3:46:38<4:09:13, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 854/1791 [3:46:54<4:08:57, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 854/1791 [3:46:54<4:08:57, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 855/1791 [3:47:09<4:08:41, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 855/1791 [3:47:09<4:08:41, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 856/1791 [3:47:26<4:08:26, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 856/1791 [3:47:26<4:08:26, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 857/1791 [3:47:42<4:08:10, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 857/1791 [3:47:42<4:08:10, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 858/1791 [3:47:58<4:07:53, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 858/1791 [3:47:58<4:07:53, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 859/1791 [3:48:14<4:07:37, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 859/1791 [3:48:14<4:07:37, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 860/1791 [3:48:29<4:07:21, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 860/1791 [3:48:29<4:07:21, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 861/1791 [3:48:44<4:07:04, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 861/1791 [3:48:44<4:07:04, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 862/1791 [3:49:00<4:06:48, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 862/1791 [3:49:00<4:06:48, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 863/1791 [3:49:16<4:06:32, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 863/1791 [3:49:16<4:06:32, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 864/1791 [3:49:33<4:06:17, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 864/1791 [3:49:33<4:06:17, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 865/1791 [3:49:48<4:06:01, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 865/1791 [3:49:48<4:06:01, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 866/1791 [3:50:04<4:05:44, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 866/1791 [3:50:04<4:05:44, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 867/1791 [3:50:20<4:05:28, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 867/1791 [3:50:20<4:05:28, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 868/1791 [3:50:35<4:05:12, 0.06it/s, v_num=bc8f] Epoch 0: 48%|████▊ | 868/1791 [3:50:35<4:05:12, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 869/1791 [3:50:51<4:04:56, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 869/1791 [3:50:51<4:04:56, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 870/1791 [3:51:06<4:04:39, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 870/1791 [3:51:06<4:04:39, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 871/1791 [3:51:22<4:04:23, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 871/1791 [3:51:22<4:04:23, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 872/1791 [3:51:39<4:04:08, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 872/1791 [3:51:39<4:04:08, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 873/1791 [3:51:55<4:03:52, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▊ | 873/1791 [3:51:55<4:03:52, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 874/1791 [3:52:11<4:03:36, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 874/1791 [3:52:11<4:03:36, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 875/1791 [3:52:26<4:03:19, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 875/1791 [3:52:26<4:03:19, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 876/1791 [3:52:42<4:03:03, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 876/1791 [3:52:42<4:03:03, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 877/1791 [3:52:57<4:02:47, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 877/1791 [3:52:57<4:02:47, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 878/1791 [3:53:13<4:02:31, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 878/1791 [3:53:13<4:02:31, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 879/1791 [3:53:29<4:02:15, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 879/1791 [3:53:29<4:02:15, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 880/1791 [3:53:46<4:02:00, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 880/1791 [3:53:46<4:02:00, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 881/1791 [3:54:02<4:01:44, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 881/1791 [3:54:02<4:01:44, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 882/1791 [3:54:17<4:01:27, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 882/1791 [3:54:17<4:01:27, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 883/1791 [3:54:32<4:01:10, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 883/1791 [3:54:32<4:01:10, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 884/1791 [3:54:47<4:00:54, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 884/1791 [3:54:47<4:00:54, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 885/1791 [3:55:03<4:00:38, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 885/1791 [3:55:03<4:00:38, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 886/1791 [3:55:19<4:00:22, 0.06it/s, v_num=bc8f] Epoch 0: 49%|████▉ | 886/1791 [3:55:19<4:00:22, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 887/1791 [3:55:35<4:00:06, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 887/1791 [3:55:35<4:00:06, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 888/1791 [3:55:51<3:59:50, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 888/1791 [3:55:51<3:59:50, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 889/1791 [3:56:07<3:59:34, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 889/1791 [3:56:07<3:59:34, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 890/1791 [3:56:23<3:59:18, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 890/1791 [3:56:23<3:59:18, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 891/1791 [3:56:39<3:59:03, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 891/1791 [3:56:39<3:59:03, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 892/1791 [3:56:55<3:58:47, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 892/1791 [3:56:55<3:58:47, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 893/1791 [3:57:11<3:58:31, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 893/1791 [3:57:11<3:58:31, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 894/1791 [3:57:27<3:58:15, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 894/1791 [3:57:27<3:58:15, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 895/1791 [3:57:43<3:57:59, 0.06it/s, v_num=bc8f] Epoch 0: 50%|████▉ | 895/1791 [3:57:43<3:57:59, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 896/1791 [3:58:01<3:57:45, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 896/1791 [3:58:01<3:57:45, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 897/1791 [3:58:17<3:57:29, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 897/1791 [3:58:17<3:57:29, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 898/1791 [3:58:33<3:57:13, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 898/1791 [3:58:33<3:57:13, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 899/1791 [3:58:48<3:56:57, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 899/1791 [3:58:48<3:56:57, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 900/1791 [3:59:04<3:56:41, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 900/1791 [3:59:04<3:56:41, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 901/1791 [3:59:20<3:56:24, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 901/1791 [3:59:20<3:56:24, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 902/1791 [3:59:35<3:56:08, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 902/1791 [3:59:35<3:56:08, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 903/1791 [3:59:51<3:55:52, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 903/1791 [3:59:51<3:55:52, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 904/1791 [4:00:09<3:55:38, 0.06it/s, v_num=bc8f] Epoch 0: 50%|█████ | 904/1791 [4:00:09<3:55:38, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 905/1791 [4:00:24<3:55:22, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 905/1791 [4:00:24<3:55:22, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 906/1791 [4:00:40<3:55:06, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 906/1791 [4:00:40<3:55:06, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 907/1791 [4:00:55<3:54:49, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 907/1791 [4:00:55<3:54:49, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 908/1791 [4:01:11<3:54:33, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 908/1791 [4:01:11<3:54:33, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 909/1791 [4:01:27<3:54:17, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 909/1791 [4:01:27<3:54:17, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 910/1791 [4:01:42<3:54:00, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 910/1791 [4:01:42<3:54:00, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 911/1791 [4:01:58<3:53:44, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 911/1791 [4:01:58<3:53:44, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 912/1791 [4:02:15<3:53:29, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 912/1791 [4:02:15<3:53:29, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 913/1791 [4:02:31<3:53:13, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 913/1791 [4:02:31<3:53:13, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 914/1791 [4:02:47<3:52:57, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 914/1791 [4:02:47<3:52:57, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 915/1791 [4:03:03<3:52:42, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 915/1791 [4:03:03<3:52:42, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 916/1791 [4:03:19<3:52:26, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 916/1791 [4:03:19<3:52:26, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 917/1791 [4:03:35<3:52:10, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████ | 917/1791 [4:03:35<3:52:10, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 918/1791 [4:03:51<3:51:54, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 918/1791 [4:03:51<3:51:54, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 919/1791 [4:04:07<3:51:38, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 919/1791 [4:04:07<3:51:38, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 920/1791 [4:04:25<3:51:24, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 920/1791 [4:04:25<3:51:24, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 921/1791 [4:04:41<3:51:08, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 921/1791 [4:04:41<3:51:08, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 922/1791 [4:04:57<3:50:52, 0.06it/s, v_num=bc8f] Epoch 0: 51%|█████▏ | 922/1791 [4:04:57<3:50:52, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 923/1791 [4:05:13<3:50:36, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 923/1791 [4:05:13<3:50:36, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 924/1791 [4:05:29<3:50:20, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 924/1791 [4:05:29<3:50:20, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 925/1791 [4:05:44<3:50:04, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 925/1791 [4:05:44<3:50:04, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 926/1791 [4:06:00<3:49:48, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 926/1791 [4:06:00<3:49:48, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 927/1791 [4:06:16<3:49:32, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 927/1791 [4:06:16<3:49:32, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 928/1791 [4:06:34<3:49:17, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 928/1791 [4:06:34<3:49:17, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 929/1791 [4:06:49<3:49:01, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 929/1791 [4:06:49<3:49:01, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 930/1791 [4:07:05<3:48:45, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 930/1791 [4:07:05<3:48:45, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 931/1791 [4:07:20<3:48:28, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 931/1791 [4:07:20<3:48:28, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 932/1791 [4:07:36<3:48:13, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 932/1791 [4:07:36<3:48:13, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 933/1791 [4:07:53<3:47:57, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 933/1791 [4:07:53<3:47:57, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 934/1791 [4:08:09<3:47:41, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 934/1791 [4:08:09<3:47:41, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 935/1791 [4:08:25<3:47:25, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 935/1791 [4:08:25<3:47:25, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 936/1791 [4:08:42<3:47:11, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 936/1791 [4:08:42<3:47:11, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 937/1791 [4:08:58<3:46:55, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 937/1791 [4:08:58<3:46:55, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 938/1791 [4:09:14<3:46:38, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 938/1791 [4:09:14<3:46:38, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 939/1791 [4:09:30<3:46:23, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 939/1791 [4:09:30<3:46:23, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 940/1791 [4:09:46<3:46:07, 0.06it/s, v_num=bc8f] Epoch 0: 52%|█████▏ | 940/1791 [4:09:46<3:46:07, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 941/1791 [4:10:01<3:45:50, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 941/1791 [4:10:01<3:45:50, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 942/1791 [4:10:17<3:45:34, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 942/1791 [4:10:17<3:45:34, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 943/1791 [4:10:32<3:45:18, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 943/1791 [4:10:32<3:45:18, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 944/1791 [4:10:49<3:45:03, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 944/1791 [4:10:49<3:45:03, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 945/1791 [4:11:04<3:44:46, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 945/1791 [4:11:04<3:44:46, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 946/1791 [4:11:20<3:44:30, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 946/1791 [4:11:20<3:44:30, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 947/1791 [4:11:36<3:44:14, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 947/1791 [4:11:36<3:44:14, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 948/1791 [4:11:52<3:43:58, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 948/1791 [4:11:52<3:43:58, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 949/1791 [4:12:07<3:43:41, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 949/1791 [4:12:07<3:43:41, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 950/1791 [4:12:23<3:43:25, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 950/1791 [4:12:23<3:43:25, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 951/1791 [4:12:38<3:43:09, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 951/1791 [4:12:38<3:43:09, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 952/1791 [4:12:55<3:42:54, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 952/1791 [4:12:55<3:42:54, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 953/1791 [4:13:11<3:42:38, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 953/1791 [4:13:11<3:42:38, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 954/1791 [4:13:26<3:42:21, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 954/1791 [4:13:26<3:42:21, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 955/1791 [4:13:42<3:42:05, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 955/1791 [4:13:42<3:42:05, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 956/1791 [4:13:58<3:41:49, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 956/1791 [4:13:58<3:41:49, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 957/1791 [4:14:13<3:41:32, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 957/1791 [4:14:13<3:41:32, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 958/1791 [4:14:28<3:41:16, 0.06it/s, v_num=bc8f] Epoch 0: 53%|█████▎ | 958/1791 [4:14:28<3:41:16, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 959/1791 [4:14:44<3:40:59, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 959/1791 [4:14:44<3:40:59, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 960/1791 [4:15:00<3:40:44, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 960/1791 [4:15:00<3:40:44, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 961/1791 [4:15:16<3:40:28, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 961/1791 [4:15:16<3:40:28, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 962/1791 [4:15:31<3:40:12, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▎ | 962/1791 [4:15:31<3:40:12, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 963/1791 [4:15:47<3:39:56, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 963/1791 [4:15:47<3:39:56, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 964/1791 [4:16:03<3:39:39, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 964/1791 [4:16:03<3:39:39, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 965/1791 [4:16:18<3:39:23, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 965/1791 [4:16:18<3:39:23, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 966/1791 [4:16:34<3:39:07, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 966/1791 [4:16:34<3:39:07, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 967/1791 [4:16:50<3:38:51, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 967/1791 [4:16:50<3:38:51, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 968/1791 [4:17:06<3:38:36, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 968/1791 [4:17:06<3:38:36, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 969/1791 [4:17:22<3:38:20, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 969/1791 [4:17:22<3:38:20, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 970/1791 [4:17:38<3:38:03, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 970/1791 [4:17:38<3:38:03, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 971/1791 [4:17:54<3:37:47, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 971/1791 [4:17:54<3:37:47, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 972/1791 [4:18:09<3:37:31, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 972/1791 [4:18:09<3:37:31, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 973/1791 [4:18:25<3:37:15, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 973/1791 [4:18:25<3:37:15, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 974/1791 [4:18:41<3:36:59, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 974/1791 [4:18:41<3:36:59, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 975/1791 [4:18:56<3:36:42, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 975/1791 [4:18:56<3:36:42, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 976/1791 [4:19:13<3:36:27, 0.06it/s, v_num=bc8f] Epoch 0: 54%|█████▍ | 976/1791 [4:19:13<3:36:27, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 977/1791 [4:19:28<3:36:11, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 977/1791 [4:19:28<3:36:11, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 978/1791 [4:19:43<3:35:54, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 978/1791 [4:19:43<3:35:54, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 979/1791 [4:19:59<3:35:38, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 979/1791 [4:19:59<3:35:38, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 980/1791 [4:20:15<3:35:22, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 980/1791 [4:20:15<3:35:22, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 981/1791 [4:20:30<3:35:05, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 981/1791 [4:20:30<3:35:05, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 982/1791 [4:20:45<3:34:49, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 982/1791 [4:20:45<3:34:49, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 983/1791 [4:21:01<3:34:32, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 983/1791 [4:21:01<3:34:32, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 984/1791 [4:21:18<3:34:17, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 984/1791 [4:21:18<3:34:17, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 985/1791 [4:21:33<3:34:01, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▍ | 985/1791 [4:21:33<3:34:01, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 986/1791 [4:21:48<3:33:44, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 986/1791 [4:21:48<3:33:44, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 987/1791 [4:22:04<3:33:29, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 987/1791 [4:22:04<3:33:29, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 988/1791 [4:22:20<3:33:13, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 988/1791 [4:22:20<3:33:13, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 989/1791 [4:22:36<3:32:57, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 989/1791 [4:22:36<3:32:57, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 990/1791 [4:22:52<3:32:41, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 990/1791 [4:22:52<3:32:41, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 991/1791 [4:23:08<3:32:25, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 991/1791 [4:23:08<3:32:25, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 992/1791 [4:23:26<3:32:10, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 992/1791 [4:23:26<3:32:10, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 993/1791 [4:23:41<3:31:54, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 993/1791 [4:23:41<3:31:54, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 994/1791 [4:23:57<3:31:38, 0.06it/s, v_num=bc8f] Epoch 0: 55%|█████▌ | 994/1791 [4:23:57<3:31:38, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 995/1791 [4:24:13<3:31:22, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 995/1791 [4:24:13<3:31:22, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 996/1791 [4:24:29<3:31:06, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 996/1791 [4:24:29<3:31:06, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 997/1791 [4:24:44<3:30:50, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 997/1791 [4:24:44<3:30:50, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 998/1791 [4:25:00<3:30:34, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 998/1791 [4:25:00<3:30:34, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 999/1791 [4:25:16<3:30:18, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 999/1791 [4:25:16<3:30:18, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1000/1791 [4:25:33<3:30:03, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1000/1791 [4:25:33<3:30:03, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1001/1791 [4:25:48<3:29:46, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1001/1791 [4:25:48<3:29:46, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1002/1791 [4:26:04<3:29:30, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1002/1791 [4:26:04<3:29:30, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1003/1791 [4:26:20<3:29:14, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1003/1791 [4:26:20<3:29:14, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1004/1791 [4:26:36<3:28:58, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1004/1791 [4:26:36<3:28:58, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1005/1791 [4:26:51<3:28:42, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1005/1791 [4:26:51<3:28:42, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1006/1791 [4:27:07<3:28:26, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1006/1791 [4:27:07<3:28:26, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1007/1791 [4:27:23<3:28:10, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▌ | 1007/1791 [4:27:23<3:28:10, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1008/1791 [4:27:40<3:27:55, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1008/1791 [4:27:40<3:27:55, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1009/1791 [4:27:55<3:27:38, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1009/1791 [4:27:55<3:27:38, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1010/1791 [4:28:11<3:27:22, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1010/1791 [4:28:11<3:27:22, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1011/1791 [4:28:26<3:27:06, 0.06it/s, v_num=bc8f] Epoch 0: 56%|█████▋ | 1011/1791 [4:28:26<3:27:06, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1012/1791 [4:28:42<3:26:50, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1012/1791 [4:28:42<3:26:50, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1013/1791 [4:28:58<3:26:34, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1013/1791 [4:28:58<3:26:34, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1014/1791 [4:29:13<3:26:18, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1014/1791 [4:29:14<3:26:18, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1015/1791 [4:29:29<3:26:02, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1015/1791 [4:29:29<3:26:02, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1016/1791 [4:29:46<3:25:47, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1016/1791 [4:29:46<3:25:47, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1017/1791 [4:30:01<3:25:30, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1017/1791 [4:30:01<3:25:30, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1018/1791 [4:30:17<3:25:14, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1018/1791 [4:30:17<3:25:14, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1019/1791 [4:30:32<3:24:58, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1019/1791 [4:30:32<3:24:58, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1020/1791 [4:30:48<3:24:41, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1020/1791 [4:30:48<3:24:41, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1021/1791 [4:31:04<3:24:25, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1021/1791 [4:31:04<3:24:25, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1022/1791 [4:31:18<3:24:09, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1022/1791 [4:31:18<3:24:09, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1023/1791 [4:31:34<3:23:52, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1023/1791 [4:31:34<3:23:52, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1024/1791 [4:31:51<3:23:37, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1024/1791 [4:31:51<3:23:37, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1025/1791 [4:32:06<3:23:21, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1025/1791 [4:32:06<3:23:21, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1026/1791 [4:32:22<3:23:04, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1026/1791 [4:32:22<3:23:04, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1027/1791 [4:32:37<3:22:48, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1027/1791 [4:32:37<3:22:48, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1028/1791 [4:32:52<3:22:32, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1028/1791 [4:32:52<3:22:32, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1029/1791 [4:33:08<3:22:15, 0.06it/s, v_num=bc8f] Epoch 0: 57%|█████▋ | 1029/1791 [4:33:08<3:22:15, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1030/1791 [4:33:24<3:22:00, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1030/1791 [4:33:24<3:22:00, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1031/1791 [4:33:40<3:21:44, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1031/1791 [4:33:40<3:21:44, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1032/1791 [4:33:57<3:21:29, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1032/1791 [4:33:57<3:21:29, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1033/1791 [4:34:13<3:21:13, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1033/1791 [4:34:13<3:21:13, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1034/1791 [4:34:29<3:20:57, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1034/1791 [4:34:29<3:20:57, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1035/1791 [4:34:45<3:20:41, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1035/1791 [4:34:45<3:20:41, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1036/1791 [4:35:00<3:20:25, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1036/1791 [4:35:00<3:20:25, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1037/1791 [4:35:16<3:20:09, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1037/1791 [4:35:16<3:20:09, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1038/1791 [4:35:32<3:19:53, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1038/1791 [4:35:32<3:19:53, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1039/1791 [4:35:48<3:19:37, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1039/1791 [4:35:48<3:19:37, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1040/1791 [4:36:05<3:19:22, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1040/1791 [4:36:05<3:19:22, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1041/1791 [4:36:20<3:19:05, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1041/1791 [4:36:20<3:19:05, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1042/1791 [4:36:36<3:18:49, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1042/1791 [4:36:36<3:18:49, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1043/1791 [4:36:51<3:18:33, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1043/1791 [4:36:51<3:18:33, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1044/1791 [4:37:07<3:18:17, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1044/1791 [4:37:07<3:18:17, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1045/1791 [4:37:23<3:18:01, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1045/1791 [4:37:23<3:18:01, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1046/1791 [4:37:38<3:17:44, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1046/1791 [4:37:38<3:17:44, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1047/1791 [4:37:53<3:17:28, 0.06it/s, v_num=bc8f] Epoch 0: 58%|█████▊ | 1047/1791 [4:37:53<3:17:28, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1048/1791 [4:38:10<3:17:13, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1048/1791 [4:38:10<3:17:13, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1049/1791 [4:38:26<3:16:56, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1049/1791 [4:38:26<3:16:56, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1050/1791 [4:38:40<3:16:40, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1050/1791 [4:38:40<3:16:40, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1051/1791 [4:38:56<3:16:23, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1051/1791 [4:38:56<3:16:23, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1052/1791 [4:39:11<3:16:07, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▊ | 1052/1791 [4:39:11<3:16:07, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1053/1791 [4:39:27<3:15:51, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1053/1791 [4:39:27<3:15:51, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1054/1791 [4:39:43<3:15:35, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1054/1791 [4:39:43<3:15:35, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1055/1791 [4:39:58<3:15:18, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1055/1791 [4:39:58<3:15:18, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1056/1791 [4:40:14<3:15:03, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1056/1791 [4:40:14<3:15:03, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1057/1791 [4:40:29<3:14:47, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1057/1791 [4:40:29<3:14:47, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1058/1791 [4:40:45<3:14:30, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1058/1791 [4:40:45<3:14:30, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1059/1791 [4:41:00<3:14:14, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1059/1791 [4:41:00<3:14:14, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1060/1791 [4:41:16<3:13:58, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1060/1791 [4:41:16<3:13:58, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1061/1791 [4:41:32<3:13:42, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1061/1791 [4:41:32<3:13:42, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1062/1791 [4:41:48<3:13:26, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1062/1791 [4:41:48<3:13:26, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1063/1791 [4:42:03<3:13:10, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1063/1791 [4:42:03<3:13:10, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1064/1791 [4:42:19<3:12:54, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1064/1791 [4:42:19<3:12:54, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1065/1791 [4:42:35<3:12:38, 0.06it/s, v_num=bc8f] Epoch 0: 59%|█████▉ | 1065/1791 [4:42:35<3:12:38, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1066/1791 [4:42:50<3:12:21, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1066/1791 [4:42:50<3:12:21, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1067/1791 [4:43:06<3:12:05, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1067/1791 [4:43:06<3:12:05, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1068/1791 [4:43:21<3:11:49, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1068/1791 [4:43:21<3:11:49, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1069/1791 [4:43:37<3:11:33, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1069/1791 [4:43:37<3:11:33, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1070/1791 [4:43:53<3:11:17, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1070/1791 [4:43:53<3:11:17, 0.06it/s, v_num=bc8f]Token indices sequence length is longer than the specified maximum sequence length for this model (1044 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 60%|█████▉ | 1071/1791 [4:44:09<3:11:01, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1071/1791 [4:44:09<3:11:01, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1072/1791 [4:44:25<3:10:46, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1072/1791 [4:44:25<3:10:46, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1073/1791 [4:44:41<3:10:29, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1073/1791 [4:44:41<3:10:29, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1074/1791 [4:44:57<3:10:13, 0.06it/s, v_num=bc8f] Epoch 0: 60%|█████▉ | 1074/1791 [4:44:57<3:10:13, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1075/1791 [4:45:12<3:09:57, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1075/1791 [4:45:12<3:09:57, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1076/1791 [4:45:28<3:09:42, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1076/1791 [4:45:28<3:09:42, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1077/1791 [4:45:44<3:09:25, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1077/1791 [4:45:44<3:09:25, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1078/1791 [4:45:59<3:09:09, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1078/1791 [4:45:59<3:09:09, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1079/1791 [4:46:15<3:08:53, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1079/1791 [4:46:15<3:08:53, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1080/1791 [4:46:32<3:08:38, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1080/1791 [4:46:32<3:08:38, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1081/1791 [4:46:48<3:08:22, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1081/1791 [4:46:48<3:08:22, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1082/1791 [4:47:03<3:08:05, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1082/1791 [4:47:03<3:08:05, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1083/1791 [4:47:18<3:07:49, 0.06it/s, v_num=bc8f] Epoch 0: 60%|██████ | 1083/1791 [4:47:18<3:07:49, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1084/1791 [4:47:34<3:07:33, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1084/1791 [4:47:34<3:07:33, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1085/1791 [4:47:49<3:07:17, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1085/1791 [4:47:49<3:07:17, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1086/1791 [4:48:05<3:07:01, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1086/1791 [4:48:05<3:07:01, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1087/1791 [4:48:20<3:06:44, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1087/1791 [4:48:20<3:06:44, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1088/1791 [4:48:37<3:06:29, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1088/1791 [4:48:37<3:06:29, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1089/1791 [4:48:52<3:06:13, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1089/1791 [4:48:52<3:06:13, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1090/1791 [4:49:08<3:05:57, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1090/1791 [4:49:08<3:05:57, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1091/1791 [4:49:24<3:05:41, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1091/1791 [4:49:24<3:05:41, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1092/1791 [4:49:40<3:05:25, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1092/1791 [4:49:40<3:05:25, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1093/1791 [4:49:55<3:05:09, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1093/1791 [4:49:55<3:05:09, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1094/1791 [4:50:11<3:04:52, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1094/1791 [4:50:11<3:04:52, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1095/1791 [4:50:26<3:04:36, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1095/1791 [4:50:26<3:04:36, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1096/1791 [4:50:43<3:04:21, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████ | 1096/1791 [4:50:43<3:04:21, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1097/1791 [4:50:58<3:04:05, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1097/1791 [4:50:58<3:04:05, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1098/1791 [4:51:14<3:03:48, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1098/1791 [4:51:14<3:03:48, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1099/1791 [4:51:30<3:03:32, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1099/1791 [4:51:30<3:03:32, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1100/1791 [4:51:46<3:03:17, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1100/1791 [4:51:46<3:03:17, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1101/1791 [4:52:01<3:03:00, 0.06it/s, v_num=bc8f] Epoch 0: 61%|██████▏ | 1101/1791 [4:52:01<3:03:00, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1102/1791 [4:52:16<3:02:44, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1102/1791 [4:52:16<3:02:44, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1103/1791 [4:52:32<3:02:28, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1103/1791 [4:52:32<3:02:28, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1104/1791 [4:52:48<3:02:12, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1104/1791 [4:52:48<3:02:12, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1105/1791 [4:53:04<3:01:56, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1105/1791 [4:53:04<3:01:56, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1106/1791 [4:53:20<3:01:40, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1106/1791 [4:53:20<3:01:40, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1107/1791 [4:53:35<3:01:24, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1107/1791 [4:53:35<3:01:24, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1108/1791 [4:53:51<3:01:08, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1108/1791 [4:53:51<3:01:08, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1109/1791 [4:54:06<3:00:52, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1109/1791 [4:54:06<3:00:52, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1110/1791 [4:54:22<3:00:36, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1110/1791 [4:54:22<3:00:36, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1111/1791 [4:54:37<3:00:19, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1111/1791 [4:54:37<3:00:19, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1112/1791 [4:54:55<3:00:04, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1112/1791 [4:54:55<3:00:04, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1113/1791 [4:55:11<2:59:49, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1113/1791 [4:55:11<2:59:49, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1114/1791 [4:55:26<2:59:32, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1114/1791 [4:55:26<2:59:32, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1115/1791 [4:55:41<2:59:16, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1115/1791 [4:55:41<2:59:16, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1116/1791 [4:55:57<2:59:00, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1116/1791 [4:55:57<2:59:00, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1117/1791 [4:56:12<2:58:44, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1117/1791 [4:56:12<2:58:44, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1118/1791 [4:56:28<2:58:27, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1118/1791 [4:56:28<2:58:27, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1119/1791 [4:56:43<2:58:11, 0.06it/s, v_num=bc8f] Epoch 0: 62%|██████▏ | 1119/1791 [4:56:43<2:58:11, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1120/1791 [4:57:01<2:57:56, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1120/1791 [4:57:01<2:57:56, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1121/1791 [4:57:16<2:57:40, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1121/1791 [4:57:16<2:57:40, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1122/1791 [4:57:31<2:57:24, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1122/1791 [4:57:31<2:57:24, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1123/1791 [4:57:47<2:57:08, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1123/1791 [4:57:47<2:57:08, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1124/1791 [4:58:02<2:56:52, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1124/1791 [4:58:02<2:56:52, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1125/1791 [4:58:17<2:56:35, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1125/1791 [4:58:17<2:56:35, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1126/1791 [4:58:33<2:56:19, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1126/1791 [4:58:33<2:56:19, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1127/1791 [4:58:48<2:56:02, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1127/1791 [4:58:48<2:56:02, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1128/1791 [4:59:05<2:55:47, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1128/1791 [4:59:05<2:55:47, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1129/1791 [4:59:21<2:55:31, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1129/1791 [4:59:21<2:55:31, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1130/1791 [4:59:37<2:55:16, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1130/1791 [4:59:37<2:55:16, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1131/1791 [4:59:53<2:55:00, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1131/1791 [4:59:53<2:55:00, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1132/1791 [5:00:08<2:54:44, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1132/1791 [5:00:08<2:54:44, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1133/1791 [5:00:23<2:54:27, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1133/1791 [5:00:23<2:54:27, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1134/1791 [5:00:39<2:54:11, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1134/1791 [5:00:39<2:54:11, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1135/1791 [5:00:54<2:53:55, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1135/1791 [5:00:54<2:53:55, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1136/1791 [5:01:12<2:53:40, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1136/1791 [5:01:12<2:53:40, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1137/1791 [5:01:27<2:53:23, 0.06it/s, v_num=bc8f] Epoch 0: 63%|██████▎ | 1137/1791 [5:01:27<2:53:23, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1138/1791 [5:01:42<2:53:07, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1138/1791 [5:01:42<2:53:07, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1139/1791 [5:01:58<2:52:51, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1139/1791 [5:01:58<2:52:51, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1140/1791 [5:02:14<2:52:35, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1140/1791 [5:02:14<2:52:35, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1141/1791 [5:02:29<2:52:19, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▎ | 1141/1791 [5:02:29<2:52:19, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1142/1791 [5:02:45<2:52:03, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1142/1791 [5:02:45<2:52:03, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1143/1791 [5:03:00<2:51:47, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1143/1791 [5:03:00<2:51:47, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1144/1791 [5:03:18<2:51:32, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1144/1791 [5:03:18<2:51:32, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1145/1791 [5:03:34<2:51:16, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1145/1791 [5:03:34<2:51:16, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1146/1791 [5:03:49<2:51:00, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1146/1791 [5:03:49<2:51:00, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1147/1791 [5:04:05<2:50:44, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1147/1791 [5:04:05<2:50:44, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1148/1791 [5:04:20<2:50:28, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1148/1791 [5:04:20<2:50:28, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1149/1791 [5:04:36<2:50:12, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1149/1791 [5:04:36<2:50:12, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1150/1791 [5:04:51<2:49:55, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1150/1791 [5:04:51<2:49:55, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1151/1791 [5:05:06<2:49:39, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1151/1791 [5:05:06<2:49:39, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1152/1791 [5:05:23<2:49:23, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1152/1791 [5:05:23<2:49:23, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1153/1791 [5:05:39<2:49:07, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1153/1791 [5:05:39<2:49:07, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1154/1791 [5:05:54<2:48:51, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1154/1791 [5:05:54<2:48:51, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1155/1791 [5:06:10<2:48:35, 0.06it/s, v_num=bc8f] Epoch 0: 64%|██████▍ | 1155/1791 [5:06:10<2:48:35, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1156/1791 [5:06:26<2:48:19, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1156/1791 [5:06:26<2:48:19, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1157/1791 [5:06:41<2:48:03, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1157/1791 [5:06:41<2:48:03, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1158/1791 [5:06:56<2:47:47, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1158/1791 [5:06:56<2:47:47, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1159/1791 [5:07:12<2:47:31, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1159/1791 [5:07:12<2:47:31, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1160/1791 [5:07:29<2:47:15, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1160/1791 [5:07:29<2:47:15, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1161/1791 [5:07:45<2:46:59, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1161/1791 [5:07:45<2:46:59, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1162/1791 [5:08:00<2:46:43, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1162/1791 [5:08:00<2:46:43, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1163/1791 [5:08:16<2:46:27, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1163/1791 [5:08:16<2:46:27, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1164/1791 [5:08:31<2:46:11, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▍ | 1164/1791 [5:08:31<2:46:11, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1165/1791 [5:08:46<2:45:55, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1165/1791 [5:08:46<2:45:55, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1166/1791 [5:09:02<2:45:39, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1166/1791 [5:09:02<2:45:39, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1167/1791 [5:09:18<2:45:23, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1167/1791 [5:09:18<2:45:23, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1168/1791 [5:09:35<2:45:07, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1168/1791 [5:09:35<2:45:07, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1169/1791 [5:09:50<2:44:51, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1169/1791 [5:09:50<2:44:51, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1170/1791 [5:10:06<2:44:35, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1170/1791 [5:10:06<2:44:35, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1171/1791 [5:10:22<2:44:19, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1171/1791 [5:10:22<2:44:19, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1172/1791 [5:10:37<2:44:03, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1172/1791 [5:10:37<2:44:03, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1173/1791 [5:10:52<2:43:47, 0.06it/s, v_num=bc8f] Epoch 0: 65%|██████▌ | 1173/1791 [5:10:52<2:43:47, 0.06it/s, v_num=bc8f]Token indices sequence length is longer than the specified maximum sequence length for this model (1078 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 66%|██████▌ | 1174/1791 [5:11:08<2:43:31, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1174/1791 [5:11:08<2:43:31, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1175/1791 [5:11:24<2:43:15, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1175/1791 [5:11:24<2:43:15, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1176/1791 [5:11:41<2:43:00, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1176/1791 [5:11:41<2:43:00, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1177/1791 [5:11:57<2:42:44, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1177/1791 [5:11:57<2:42:44, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1178/1791 [5:12:13<2:42:28, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1178/1791 [5:12:13<2:42:28, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1179/1791 [5:12:28<2:42:12, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1179/1791 [5:12:28<2:42:12, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1180/1791 [5:12:44<2:41:56, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1180/1791 [5:12:44<2:41:56, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1181/1791 [5:13:00<2:41:40, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1181/1791 [5:13:00<2:41:40, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1182/1791 [5:13:15<2:41:24, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1182/1791 [5:13:15<2:41:24, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1183/1791 [5:13:31<2:41:07, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1183/1791 [5:13:31<2:41:07, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1184/1791 [5:13:48<2:40:52, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1184/1791 [5:13:48<2:40:52, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1185/1791 [5:14:03<2:40:36, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1185/1791 [5:14:03<2:40:36, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1186/1791 [5:14:19<2:40:20, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▌ | 1186/1791 [5:14:19<2:40:20, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1187/1791 [5:14:34<2:40:04, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1187/1791 [5:14:34<2:40:04, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1188/1791 [5:14:50<2:39:48, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1188/1791 [5:14:50<2:39:48, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1189/1791 [5:15:05<2:39:32, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1189/1791 [5:15:05<2:39:32, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1190/1791 [5:15:21<2:39:16, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1190/1791 [5:15:21<2:39:16, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1191/1791 [5:15:36<2:38:59, 0.06it/s, v_num=bc8f] Epoch 0: 66%|██████▋ | 1191/1791 [5:15:36<2:38:59, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1192/1791 [5:15:53<2:38:44, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1192/1791 [5:15:53<2:38:44, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1193/1791 [5:16:09<2:38:28, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1193/1791 [5:16:09<2:38:28, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1194/1791 [5:16:24<2:38:12, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1194/1791 [5:16:24<2:38:12, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1195/1791 [5:16:39<2:37:55, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1195/1791 [5:16:39<2:37:55, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1196/1791 [5:16:53<2:37:39, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1196/1791 [5:16:53<2:37:39, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1197/1791 [5:17:09<2:37:23, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1197/1791 [5:17:09<2:37:23, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1198/1791 [5:17:24<2:37:07, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1198/1791 [5:17:24<2:37:07, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1199/1791 [5:17:41<2:36:51, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1199/1791 [5:17:41<2:36:51, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1200/1791 [5:17:57<2:36:35, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1200/1791 [5:17:57<2:36:35, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1201/1791 [5:18:13<2:36:19, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1201/1791 [5:18:13<2:36:19, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1202/1791 [5:18:28<2:36:03, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1202/1791 [5:18:28<2:36:03, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1203/1791 [5:18:44<2:35:47, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1203/1791 [5:18:44<2:35:47, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1204/1791 [5:19:00<2:35:31, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1204/1791 [5:19:00<2:35:31, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1205/1791 [5:19:15<2:35:15, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1205/1791 [5:19:15<2:35:15, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1206/1791 [5:19:30<2:34:59, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1206/1791 [5:19:30<2:34:59, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1207/1791 [5:19:46<2:34:43, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1207/1791 [5:19:46<2:34:43, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1208/1791 [5:20:04<2:34:28, 0.06it/s, v_num=bc8f] Epoch 0: 67%|██████▋ | 1208/1791 [5:20:04<2:34:28, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1209/1791 [5:20:20<2:34:12, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1209/1791 [5:20:20<2:34:12, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1210/1791 [5:20:35<2:33:56, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1210/1791 [5:20:35<2:33:56, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1211/1791 [5:20:51<2:33:40, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1211/1791 [5:20:51<2:33:40, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1212/1791 [5:21:07<2:33:24, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1212/1791 [5:21:07<2:33:24, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1213/1791 [5:21:23<2:33:08, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1213/1791 [5:21:23<2:33:08, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1214/1791 [5:21:38<2:32:52, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1214/1791 [5:21:38<2:32:52, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1215/1791 [5:21:54<2:32:36, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1215/1791 [5:21:54<2:32:36, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1216/1791 [5:22:11<2:32:21, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1216/1791 [5:22:11<2:32:21, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1217/1791 [5:22:27<2:32:05, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1217/1791 [5:22:27<2:32:05, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1218/1791 [5:22:42<2:31:49, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1218/1791 [5:22:42<2:31:49, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1219/1791 [5:22:58<2:31:33, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1219/1791 [5:22:58<2:31:33, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1220/1791 [5:23:14<2:31:17, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1220/1791 [5:23:14<2:31:17, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1221/1791 [5:23:30<2:31:01, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1221/1791 [5:23:30<2:31:01, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1222/1791 [5:23:46<2:30:45, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1222/1791 [5:23:46<2:30:45, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1223/1791 [5:24:01<2:30:29, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1223/1791 [5:24:01<2:30:29, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1224/1791 [5:24:18<2:30:13, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1224/1791 [5:24:18<2:30:13, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1225/1791 [5:24:34<2:29:57, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1225/1791 [5:24:34<2:29:57, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1226/1791 [5:24:49<2:29:41, 0.06it/s, v_num=bc8f] Epoch 0: 68%|██████▊ | 1226/1791 [5:24:49<2:29:41, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1227/1791 [5:25:05<2:29:25, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1227/1791 [5:25:05<2:29:25, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1228/1791 [5:25:20<2:29:09, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1228/1791 [5:25:20<2:29:09, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1229/1791 [5:25:36<2:28:53, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1229/1791 [5:25:36<2:28:53, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1230/1791 [5:25:51<2:28:37, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1230/1791 [5:25:51<2:28:37, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1231/1791 [5:26:07<2:28:21, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▊ | 1231/1791 [5:26:07<2:28:21, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1232/1791 [5:26:24<2:28:06, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1232/1791 [5:26:24<2:28:06, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1233/1791 [5:26:39<2:27:49, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1233/1791 [5:26:39<2:27:49, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1234/1791 [5:26:55<2:27:33, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1234/1791 [5:26:55<2:27:33, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1235/1791 [5:27:10<2:27:17, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1235/1791 [5:27:10<2:27:17, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1236/1791 [5:27:26<2:27:01, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1236/1791 [5:27:26<2:27:01, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1237/1791 [5:27:41<2:26:45, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1237/1791 [5:27:41<2:26:45, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1238/1791 [5:27:57<2:26:29, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1238/1791 [5:27:57<2:26:29, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1239/1791 [5:28:13<2:26:13, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1239/1791 [5:28:13<2:26:13, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1240/1791 [5:28:30<2:25:58, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1240/1791 [5:28:30<2:25:58, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1241/1791 [5:28:45<2:25:42, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1241/1791 [5:28:45<2:25:42, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1242/1791 [5:29:01<2:25:26, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1242/1791 [5:29:01<2:25:26, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1243/1791 [5:29:17<2:25:10, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1243/1791 [5:29:17<2:25:10, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1244/1791 [5:29:32<2:24:54, 0.06it/s, v_num=bc8f] Epoch 0: 69%|██████▉ | 1244/1791 [5:29:32<2:24:54, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1245/1791 [5:29:48<2:24:38, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1245/1791 [5:29:48<2:24:38, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1246/1791 [5:30:03<2:24:22, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1246/1791 [5:30:03<2:24:22, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1247/1791 [5:30:19<2:24:06, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1247/1791 [5:30:19<2:24:06, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1248/1791 [5:30:35<2:23:50, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1248/1791 [5:30:35<2:23:50, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1249/1791 [5:30:51<2:23:34, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1249/1791 [5:30:51<2:23:34, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1250/1791 [5:31:07<2:23:18, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1250/1791 [5:31:07<2:23:18, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1251/1791 [5:31:23<2:23:02, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1251/1791 [5:31:23<2:23:02, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1252/1791 [5:31:38<2:22:46, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1252/1791 [5:31:38<2:22:46, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1253/1791 [5:31:53<2:22:30, 0.06it/s, v_num=bc8f] Epoch 0: 70%|██████▉ | 1253/1791 [5:31:53<2:22:30, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1254/1791 [5:32:09<2:22:14, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1254/1791 [5:32:09<2:22:14, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1255/1791 [5:32:24<2:21:58, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1255/1791 [5:32:24<2:21:58, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1256/1791 [5:32:41<2:21:42, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1256/1791 [5:32:41<2:21:42, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1257/1791 [5:32:57<2:21:26, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1257/1791 [5:32:57<2:21:26, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1258/1791 [5:33:12<2:21:10, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1258/1791 [5:33:12<2:21:10, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1259/1791 [5:33:29<2:20:55, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1259/1791 [5:33:29<2:20:55, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1260/1791 [5:33:44<2:20:39, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1260/1791 [5:33:44<2:20:39, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1261/1791 [5:34:00<2:20:23, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1261/1791 [5:34:00<2:20:23, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1262/1791 [5:34:16<2:20:07, 0.06it/s, v_num=bc8f] Epoch 0: 70%|███████ | 1262/1791 [5:34:16<2:20:07, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1263/1791 [5:34:31<2:19:51, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1263/1791 [5:34:31<2:19:51, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1264/1791 [5:34:49<2:19:35, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1264/1791 [5:34:49<2:19:35, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1265/1791 [5:35:05<2:19:19, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1265/1791 [5:35:05<2:19:19, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1266/1791 [5:35:20<2:19:03, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1266/1791 [5:35:20<2:19:03, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1267/1791 [5:35:36<2:18:47, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1267/1791 [5:35:36<2:18:47, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1268/1791 [5:35:51<2:18:31, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1268/1791 [5:35:51<2:18:31, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1269/1791 [5:36:06<2:18:15, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1269/1791 [5:36:06<2:18:15, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1270/1791 [5:36:21<2:17:59, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1270/1791 [5:36:21<2:17:59, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1271/1791 [5:36:36<2:17:42, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1271/1791 [5:36:36<2:17:42, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1272/1791 [5:36:52<2:17:27, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1272/1791 [5:36:52<2:17:27, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1273/1791 [5:37:08<2:17:11, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1273/1791 [5:37:08<2:17:11, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1274/1791 [5:37:23<2:16:55, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1274/1791 [5:37:23<2:16:55, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1275/1791 [5:37:39<2:16:39, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1275/1791 [5:37:39<2:16:39, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1276/1791 [5:37:54<2:16:23, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████ | 1276/1791 [5:37:54<2:16:23, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1277/1791 [5:38:10<2:16:07, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1277/1791 [5:38:10<2:16:07, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1278/1791 [5:38:26<2:15:51, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1278/1791 [5:38:26<2:15:51, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1279/1791 [5:38:42<2:15:35, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1279/1791 [5:38:42<2:15:35, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1280/1791 [5:38:58<2:15:19, 0.06it/s, v_num=bc8f] Epoch 0: 71%|███████▏ | 1280/1791 [5:38:58<2:15:19, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1281/1791 [5:39:14<2:15:03, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1281/1791 [5:39:14<2:15:03, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1282/1791 [5:39:30<2:14:47, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1282/1791 [5:39:30<2:14:47, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1283/1791 [5:39:45<2:14:31, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1283/1791 [5:39:45<2:14:31, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1284/1791 [5:40:01<2:14:15, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1284/1791 [5:40:01<2:14:15, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1285/1791 [5:40:17<2:13:59, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1285/1791 [5:40:17<2:13:59, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1286/1791 [5:40:32<2:13:43, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1286/1791 [5:40:32<2:13:43, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1287/1791 [5:40:48<2:13:27, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1287/1791 [5:40:48<2:13:27, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1288/1791 [5:41:05<2:13:12, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1288/1791 [5:41:05<2:13:12, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1289/1791 [5:41:21<2:12:56, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1289/1791 [5:41:21<2:12:56, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1290/1791 [5:41:36<2:12:40, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1290/1791 [5:41:36<2:12:40, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1291/1791 [5:41:52<2:12:24, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1291/1791 [5:41:52<2:12:24, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1292/1791 [5:42:08<2:12:08, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1292/1791 [5:42:08<2:12:08, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1293/1791 [5:42:23<2:11:52, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1293/1791 [5:42:23<2:11:52, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1294/1791 [5:42:38<2:11:36, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1294/1791 [5:42:38<2:11:36, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1295/1791 [5:42:54<2:11:20, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1295/1791 [5:42:54<2:11:20, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1296/1791 [5:43:11<2:11:04, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1296/1791 [5:43:11<2:11:04, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1297/1791 [5:43:26<2:10:48, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1297/1791 [5:43:26<2:10:48, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1298/1791 [5:43:42<2:10:32, 0.06it/s, v_num=bc8f] Epoch 0: 72%|███████▏ | 1298/1791 [5:43:42<2:10:32, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1299/1791 [5:43:58<2:10:16, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1299/1791 [5:43:58<2:10:16, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1300/1791 [5:44:14<2:10:00, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1300/1791 [5:44:14<2:10:00, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1301/1791 [5:44:30<2:09:45, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1301/1791 [5:44:30<2:09:45, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1302/1791 [5:44:46<2:09:29, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1302/1791 [5:44:46<2:09:29, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1303/1791 [5:45:02<2:09:13, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1303/1791 [5:45:02<2:09:13, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1304/1791 [5:45:19<2:08:57, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1304/1791 [5:45:19<2:08:57, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1305/1791 [5:45:35<2:08:42, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1305/1791 [5:45:35<2:08:42, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1306/1791 [5:45:50<2:08:26, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1306/1791 [5:45:50<2:08:26, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1307/1791 [5:46:05<2:08:09, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1307/1791 [5:46:05<2:08:09, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1308/1791 [5:46:21<2:07:54, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1308/1791 [5:46:21<2:07:54, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1309/1791 [5:46:37<2:07:38, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1309/1791 [5:46:37<2:07:38, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1310/1791 [5:46:53<2:07:22, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1310/1791 [5:46:53<2:07:22, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1311/1791 [5:47:08<2:07:06, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1311/1791 [5:47:08<2:07:06, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1312/1791 [5:47:26<2:06:50, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1312/1791 [5:47:26<2:06:50, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1313/1791 [5:47:41<2:06:34, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1313/1791 [5:47:41<2:06:34, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1314/1791 [5:47:57<2:06:18, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1314/1791 [5:47:57<2:06:18, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1315/1791 [5:48:13<2:06:02, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1315/1791 [5:48:13<2:06:02, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1316/1791 [5:48:28<2:05:46, 0.06it/s, v_num=bc8f] Epoch 0: 73%|███████▎ | 1316/1791 [5:48:28<2:05:46, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1317/1791 [5:48:44<2:05:31, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1317/1791 [5:48:44<2:05:31, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1318/1791 [5:49:00<2:05:15, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1318/1791 [5:49:00<2:05:15, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1319/1791 [5:49:15<2:04:58, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1319/1791 [5:49:15<2:04:58, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1320/1791 [5:49:32<2:04:43, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▎ | 1320/1791 [5:49:32<2:04:43, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1321/1791 [5:49:49<2:04:27, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1321/1791 [5:49:49<2:04:27, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1322/1791 [5:50:04<2:04:11, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1322/1791 [5:50:04<2:04:11, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1323/1791 [5:50:19<2:03:55, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1323/1791 [5:50:19<2:03:55, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1324/1791 [5:50:34<2:03:39, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1324/1791 [5:50:34<2:03:39, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1325/1791 [5:50:50<2:03:23, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1325/1791 [5:50:50<2:03:23, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1326/1791 [5:51:05<2:03:07, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1326/1791 [5:51:05<2:03:07, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1327/1791 [5:51:20<2:02:51, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1327/1791 [5:51:20<2:02:51, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1328/1791 [5:51:38<2:02:35, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1328/1791 [5:51:38<2:02:35, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1329/1791 [5:51:53<2:02:19, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1329/1791 [5:51:53<2:02:19, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1330/1791 [5:52:09<2:02:03, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1330/1791 [5:52:09<2:02:03, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1331/1791 [5:52:24<2:01:47, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1331/1791 [5:52:24<2:01:47, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1332/1791 [5:52:40<2:01:31, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1332/1791 [5:52:40<2:01:31, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1333/1791 [5:52:56<2:01:15, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1333/1791 [5:52:56<2:01:15, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1334/1791 [5:53:11<2:00:59, 0.06it/s, v_num=bc8f] Epoch 0: 74%|███████▍ | 1334/1791 [5:53:11<2:00:59, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1335/1791 [5:53:27<2:00:43, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1335/1791 [5:53:27<2:00:43, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1336/1791 [5:53:45<2:00:28, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1336/1791 [5:53:45<2:00:28, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1337/1791 [5:54:00<2:00:12, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1337/1791 [5:54:00<2:00:12, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1338/1791 [5:54:15<1:59:56, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1338/1791 [5:54:15<1:59:56, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1339/1791 [5:54:31<1:59:40, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1339/1791 [5:54:31<1:59:40, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1340/1791 [5:54:46<1:59:24, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1340/1791 [5:54:46<1:59:24, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1341/1791 [5:55:01<1:59:08, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1341/1791 [5:55:01<1:59:08, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1342/1791 [5:55:17<1:58:52, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1342/1791 [5:55:17<1:58:52, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1343/1791 [5:55:32<1:58:36, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▍ | 1343/1791 [5:55:32<1:58:36, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1344/1791 [5:55:49<1:58:20, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1344/1791 [5:55:49<1:58:20, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1345/1791 [5:56:04<1:58:04, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1345/1791 [5:56:04<1:58:04, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1346/1791 [5:56:20<1:57:48, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1346/1791 [5:56:20<1:57:48, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1347/1791 [5:56:36<1:57:32, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1347/1791 [5:56:36<1:57:32, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1348/1791 [5:56:52<1:57:16, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1348/1791 [5:56:52<1:57:16, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1349/1791 [5:57:07<1:57:00, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1349/1791 [5:57:07<1:57:00, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1350/1791 [5:57:23<1:56:44, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1350/1791 [5:57:23<1:56:44, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1351/1791 [5:57:39<1:56:28, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1351/1791 [5:57:39<1:56:28, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1352/1791 [5:57:56<1:56:13, 0.06it/s, v_num=bc8f] Epoch 0: 75%|███████▌ | 1352/1791 [5:57:56<1:56:13, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1353/1791 [5:58:11<1:55:57, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1353/1791 [5:58:11<1:55:57, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1354/1791 [5:58:26<1:55:41, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1354/1791 [5:58:26<1:55:41, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1355/1791 [5:58:42<1:55:25, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1355/1791 [5:58:42<1:55:25, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1356/1791 [5:58:57<1:55:09, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1356/1791 [5:58:57<1:55:09, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1357/1791 [5:59:13<1:54:53, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1357/1791 [5:59:13<1:54:53, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1358/1791 [5:59:29<1:54:37, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1358/1791 [5:59:29<1:54:37, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1359/1791 [5:59:45<1:54:21, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1359/1791 [5:59:45<1:54:21, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1360/1791 [6:00:02<1:54:06, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1360/1791 [6:00:02<1:54:06, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1361/1791 [6:00:18<1:53:50, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1361/1791 [6:00:18<1:53:50, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1362/1791 [6:00:34<1:53:34, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1362/1791 [6:00:34<1:53:34, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1363/1791 [6:00:49<1:53:18, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1363/1791 [6:00:49<1:53:18, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1364/1791 [6:01:04<1:53:02, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1364/1791 [6:01:04<1:53:02, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1365/1791 [6:01:19<1:52:46, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▌ | 1365/1791 [6:01:19<1:52:46, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1366/1791 [6:01:34<1:52:29, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1366/1791 [6:01:34<1:52:29, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1367/1791 [6:01:50<1:52:13, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1367/1791 [6:01:50<1:52:13, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1368/1791 [6:02:07<1:51:58, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1368/1791 [6:02:07<1:51:58, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1369/1791 [6:02:22<1:51:42, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1369/1791 [6:02:22<1:51:42, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1370/1791 [6:02:37<1:51:25, 0.06it/s, v_num=bc8f] Epoch 0: 76%|███████▋ | 1370/1791 [6:02:37<1:51:25, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1371/1791 [6:02:52<1:51:10, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1371/1791 [6:02:52<1:51:10, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1372/1791 [6:03:08<1:50:54, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1372/1791 [6:03:08<1:50:54, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1373/1791 [6:03:24<1:50:38, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1373/1791 [6:03:24<1:50:38, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1374/1791 [6:03:39<1:50:22, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1374/1791 [6:03:39<1:50:22, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1375/1791 [6:03:55<1:50:06, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1375/1791 [6:03:55<1:50:06, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1376/1791 [6:04:12<1:49:50, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1376/1791 [6:04:12<1:49:50, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1377/1791 [6:04:28<1:49:34, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1377/1791 [6:04:28<1:49:34, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1378/1791 [6:04:43<1:49:18, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1378/1791 [6:04:43<1:49:18, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1379/1791 [6:04:59<1:49:02, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1379/1791 [6:04:59<1:49:02, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1380/1791 [6:05:14<1:48:46, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1380/1791 [6:05:14<1:48:46, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1381/1791 [6:05:29<1:48:30, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1381/1791 [6:05:29<1:48:30, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1382/1791 [6:05:44<1:48:14, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1382/1791 [6:05:44<1:48:14, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1383/1791 [6:05:59<1:47:58, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1383/1791 [6:05:59<1:47:58, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1384/1791 [6:06:16<1:47:42, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1384/1791 [6:06:16<1:47:42, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1385/1791 [6:06:32<1:47:26, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1385/1791 [6:06:32<1:47:26, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1386/1791 [6:06:47<1:47:10, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1386/1791 [6:06:47<1:47:10, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1387/1791 [6:07:02<1:46:54, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1387/1791 [6:07:02<1:46:54, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1388/1791 [6:07:17<1:46:38, 0.06it/s, v_num=bc8f] Epoch 0: 77%|███████▋ | 1388/1791 [6:07:17<1:46:38, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1389/1791 [6:07:33<1:46:22, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1389/1791 [6:07:33<1:46:22, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1390/1791 [6:07:49<1:46:06, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1390/1791 [6:07:49<1:46:06, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1391/1791 [6:08:04<1:45:50, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1391/1791 [6:08:04<1:45:50, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1392/1791 [6:08:20<1:45:34, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1392/1791 [6:08:20<1:45:34, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1393/1791 [6:08:35<1:45:18, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1393/1791 [6:08:35<1:45:18, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1394/1791 [6:08:50<1:45:02, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1394/1791 [6:08:50<1:45:02, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1395/1791 [6:09:05<1:44:46, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1395/1791 [6:09:05<1:44:46, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1396/1791 [6:09:20<1:44:30, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1396/1791 [6:09:20<1:44:30, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1397/1791 [6:09:35<1:44:14, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1397/1791 [6:09:35<1:44:14, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1398/1791 [6:09:50<1:43:58, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1398/1791 [6:09:50<1:43:58, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1399/1791 [6:10:05<1:43:42, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1399/1791 [6:10:05<1:43:42, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1400/1791 [6:10:22<1:43:26, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1400/1791 [6:10:22<1:43:26, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1401/1791 [6:10:37<1:43:10, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1401/1791 [6:10:37<1:43:10, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1402/1791 [6:10:52<1:42:54, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1402/1791 [6:10:52<1:42:54, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1403/1791 [6:11:08<1:42:38, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1403/1791 [6:11:08<1:42:38, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1404/1791 [6:11:23<1:42:22, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1404/1791 [6:11:23<1:42:22, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1405/1791 [6:11:38<1:42:06, 0.06it/s, v_num=bc8f] Epoch 0: 78%|███████▊ | 1405/1791 [6:11:38<1:42:06, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1406/1791 [6:11:54<1:41:50, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1406/1791 [6:11:54<1:41:50, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1407/1791 [6:12:09<1:41:34, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1407/1791 [6:12:09<1:41:34, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1408/1791 [6:12:26<1:41:18, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1408/1791 [6:12:26<1:41:18, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1409/1791 [6:12:42<1:41:02, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1409/1791 [6:12:42<1:41:02, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1410/1791 [6:12:58<1:40:46, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▊ | 1410/1791 [6:12:58<1:40:46, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1411/1791 [6:13:13<1:40:30, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1411/1791 [6:13:13<1:40:30, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1412/1791 [6:13:29<1:40:15, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1412/1791 [6:13:29<1:40:15, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1413/1791 [6:13:44<1:39:59, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1413/1791 [6:13:44<1:39:59, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1414/1791 [6:14:00<1:39:43, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1414/1791 [6:14:00<1:39:43, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1415/1791 [6:14:15<1:39:27, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1415/1791 [6:14:15<1:39:27, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1416/1791 [6:14:32<1:39:11, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1416/1791 [6:14:32<1:39:11, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1417/1791 [6:14:48<1:38:55, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1417/1791 [6:14:48<1:38:55, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1418/1791 [6:15:03<1:38:39, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1418/1791 [6:15:03<1:38:39, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1419/1791 [6:15:18<1:38:23, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1419/1791 [6:15:18<1:38:23, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1420/1791 [6:15:34<1:38:07, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1420/1791 [6:15:34<1:38:07, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1421/1791 [6:15:49<1:37:51, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1421/1791 [6:15:49<1:37:51, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1422/1791 [6:16:04<1:37:35, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1422/1791 [6:16:04<1:37:35, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1423/1791 [6:16:20<1:37:19, 0.06it/s, v_num=bc8f] Epoch 0: 79%|███████▉ | 1423/1791 [6:16:20<1:37:19, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1424/1791 [6:16:36<1:37:03, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1424/1791 [6:16:36<1:37:03, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1425/1791 [6:16:51<1:36:47, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1425/1791 [6:16:51<1:36:47, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1426/1791 [6:17:06<1:36:31, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1426/1791 [6:17:06<1:36:31, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1427/1791 [6:17:21<1:36:15, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1427/1791 [6:17:21<1:36:15, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1428/1791 [6:17:36<1:35:59, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1428/1791 [6:17:36<1:35:59, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1429/1791 [6:17:52<1:35:43, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1429/1791 [6:17:52<1:35:43, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1430/1791 [6:18:07<1:35:27, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1430/1791 [6:18:07<1:35:27, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1431/1791 [6:18:22<1:35:11, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1431/1791 [6:18:22<1:35:11, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1432/1791 [6:18:39<1:34:55, 0.06it/s, v_num=bc8f] Epoch 0: 80%|███████▉ | 1432/1791 [6:18:39<1:34:55, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1433/1791 [6:18:54<1:34:39, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1433/1791 [6:18:54<1:34:39, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1434/1791 [6:19:10<1:34:23, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1434/1791 [6:19:10<1:34:23, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1435/1791 [6:19:25<1:34:07, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1435/1791 [6:19:25<1:34:07, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1436/1791 [6:19:40<1:33:51, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1436/1791 [6:19:40<1:33:51, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1437/1791 [6:19:56<1:33:35, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1437/1791 [6:19:56<1:33:35, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1438/1791 [6:20:10<1:33:19, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1438/1791 [6:20:10<1:33:19, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1439/1791 [6:20:26<1:33:03, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1439/1791 [6:20:26<1:33:03, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1440/1791 [6:20:42<1:32:47, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1440/1791 [6:20:42<1:32:47, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1441/1791 [6:20:58<1:32:32, 0.06it/s, v_num=bc8f] Epoch 0: 80%|████████ | 1441/1791 [6:20:58<1:32:32, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1442/1791 [6:21:13<1:32:15, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1442/1791 [6:21:13<1:32:15, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1443/1791 [6:21:28<1:31:59, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1443/1791 [6:21:28<1:31:59, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1444/1791 [6:21:44<1:31:43, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1444/1791 [6:21:44<1:31:43, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1445/1791 [6:21:58<1:31:27, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1445/1791 [6:21:58<1:31:27, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1446/1791 [6:22:13<1:31:11, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1446/1791 [6:22:13<1:31:11, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1447/1791 [6:22:29<1:30:55, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1447/1791 [6:22:29<1:30:55, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1448/1791 [6:22:45<1:30:39, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1448/1791 [6:22:45<1:30:39, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1449/1791 [6:23:00<1:30:24, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1449/1791 [6:23:00<1:30:24, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1450/1791 [6:23:16<1:30:08, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1450/1791 [6:23:16<1:30:08, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1451/1791 [6:23:32<1:29:52, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1451/1791 [6:23:32<1:29:52, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1452/1791 [6:23:46<1:29:36, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1452/1791 [6:23:46<1:29:36, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1453/1791 [6:24:02<1:29:20, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1453/1791 [6:24:02<1:29:20, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1454/1791 [6:24:17<1:29:04, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1454/1791 [6:24:17<1:29:04, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1455/1791 [6:24:32<1:28:48, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████ | 1455/1791 [6:24:32<1:28:48, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1456/1791 [6:24:48<1:28:32, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1456/1791 [6:24:48<1:28:32, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1457/1791 [6:25:04<1:28:16, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1457/1791 [6:25:04<1:28:16, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1458/1791 [6:25:19<1:28:00, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1458/1791 [6:25:19<1:28:00, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1459/1791 [6:25:34<1:27:44, 0.06it/s, v_num=bc8f] Epoch 0: 81%|████████▏ | 1459/1791 [6:25:34<1:27:44, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1460/1791 [6:25:50<1:27:28, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1460/1791 [6:25:50<1:27:28, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1461/1791 [6:26:05<1:27:12, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1461/1791 [6:26:05<1:27:12, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1462/1791 [6:26:20<1:26:56, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1462/1791 [6:26:20<1:26:56, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1463/1791 [6:26:36<1:26:40, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1463/1791 [6:26:36<1:26:40, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1464/1791 [6:26:52<1:26:24, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1464/1791 [6:26:52<1:26:24, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1465/1791 [6:27:08<1:26:08, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1465/1791 [6:27:08<1:26:08, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1466/1791 [6:27:23<1:25:52, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1466/1791 [6:27:23<1:25:52, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1467/1791 [6:27:39<1:25:36, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1467/1791 [6:27:39<1:25:36, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1468/1791 [6:27:53<1:25:20, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1468/1791 [6:27:53<1:25:20, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1469/1791 [6:28:09<1:25:04, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1469/1791 [6:28:09<1:25:04, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1470/1791 [6:28:24<1:24:49, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1470/1791 [6:28:24<1:24:49, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1471/1791 [6:28:40<1:24:33, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1471/1791 [6:28:40<1:24:33, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1472/1791 [6:28:56<1:24:17, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1472/1791 [6:28:56<1:24:17, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1473/1791 [6:29:11<1:24:01, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1473/1791 [6:29:11<1:24:01, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1474/1791 [6:29:27<1:23:45, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1474/1791 [6:29:27<1:23:45, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1475/1791 [6:29:42<1:23:29, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1475/1791 [6:29:42<1:23:29, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1476/1791 [6:29:57<1:23:13, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1476/1791 [6:29:57<1:23:13, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1477/1791 [6:30:12<1:22:57, 0.06it/s, v_num=bc8f] Epoch 0: 82%|████████▏ | 1477/1791 [6:30:12<1:22:57, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1478/1791 [6:30:28<1:22:41, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1478/1791 [6:30:28<1:22:41, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1479/1791 [6:30:43<1:22:25, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1479/1791 [6:30:43<1:22:25, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1480/1791 [6:31:00<1:22:09, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1480/1791 [6:31:00<1:22:09, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1481/1791 [6:31:15<1:21:53, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1481/1791 [6:31:15<1:21:53, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1482/1791 [6:31:31<1:21:37, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1482/1791 [6:31:31<1:21:37, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1483/1791 [6:31:46<1:21:22, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1483/1791 [6:31:46<1:21:22, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1484/1791 [6:32:02<1:21:06, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1484/1791 [6:32:02<1:21:06, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1485/1791 [6:32:17<1:20:50, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1485/1791 [6:32:17<1:20:50, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1486/1791 [6:32:33<1:20:34, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1486/1791 [6:32:33<1:20:34, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1487/1791 [6:32:48<1:20:18, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1487/1791 [6:32:48<1:20:18, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1488/1791 [6:33:05<1:20:02, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1488/1791 [6:33:05<1:20:02, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1489/1791 [6:33:20<1:19:46, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1489/1791 [6:33:20<1:19:46, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1490/1791 [6:33:36<1:19:30, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1490/1791 [6:33:36<1:19:30, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1491/1791 [6:33:50<1:19:14, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1491/1791 [6:33:50<1:19:14, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1492/1791 [6:34:06<1:18:58, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1492/1791 [6:34:06<1:18:58, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1493/1791 [6:34:21<1:18:42, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1493/1791 [6:34:21<1:18:42, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1494/1791 [6:34:37<1:18:26, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1494/1791 [6:34:37<1:18:26, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1495/1791 [6:34:52<1:18:10, 0.06it/s, v_num=bc8f] Epoch 0: 83%|████████▎ | 1495/1791 [6:34:52<1:18:10, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1496/1791 [6:35:09<1:17:55, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1496/1791 [6:35:09<1:17:55, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1497/1791 [6:35:25<1:17:39, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1497/1791 [6:35:25<1:17:39, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1498/1791 [6:35:40<1:17:23, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1498/1791 [6:35:40<1:17:23, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1499/1791 [6:35:55<1:17:07, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▎ | 1499/1791 [6:35:55<1:17:07, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1500/1791 [6:36:10<1:16:51, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1500/1791 [6:36:10<1:16:51, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1501/1791 [6:36:26<1:16:35, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1501/1791 [6:36:26<1:16:35, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1502/1791 [6:36:41<1:16:19, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1502/1791 [6:36:41<1:16:19, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1503/1791 [6:36:56<1:16:03, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1503/1791 [6:36:56<1:16:03, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1504/1791 [6:37:13<1:15:48, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1504/1791 [6:37:13<1:15:48, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1505/1791 [6:37:29<1:15:32, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1505/1791 [6:37:29<1:15:32, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1506/1791 [6:37:44<1:15:16, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1506/1791 [6:37:44<1:15:16, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1507/1791 [6:37:59<1:15:00, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1507/1791 [6:37:59<1:15:00, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1508/1791 [6:38:14<1:14:44, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1508/1791 [6:38:14<1:14:44, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1509/1791 [6:38:30<1:14:28, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1509/1791 [6:38:30<1:14:28, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1510/1791 [6:38:45<1:14:12, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1510/1791 [6:38:45<1:14:12, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1511/1791 [6:39:01<1:13:56, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1511/1791 [6:39:01<1:13:56, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1512/1791 [6:39:17<1:13:40, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1512/1791 [6:39:17<1:13:40, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1513/1791 [6:39:33<1:13:24, 0.06it/s, v_num=bc8f] Epoch 0: 84%|████████▍ | 1513/1791 [6:39:33<1:13:24, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1514/1791 [6:39:48<1:13:08, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1514/1791 [6:39:48<1:13:08, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1515/1791 [6:40:04<1:12:53, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1515/1791 [6:40:04<1:12:53, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1516/1791 [6:40:19<1:12:37, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1516/1791 [6:40:19<1:12:37, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1517/1791 [6:40:34<1:12:21, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1517/1791 [6:40:34<1:12:21, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1518/1791 [6:40:49<1:12:05, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1518/1791 [6:40:49<1:12:05, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1519/1791 [6:41:05<1:11:49, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1519/1791 [6:41:05<1:11:49, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1520/1791 [6:41:21<1:11:33, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1520/1791 [6:41:21<1:11:33, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1521/1791 [6:41:36<1:11:17, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1521/1791 [6:41:36<1:11:17, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1522/1791 [6:41:52<1:11:01, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▍ | 1522/1791 [6:41:52<1:11:01, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1523/1791 [6:42:07<1:10:45, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1523/1791 [6:42:07<1:10:45, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1524/1791 [6:42:22<1:10:29, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1524/1791 [6:42:22<1:10:29, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1525/1791 [6:42:37<1:10:13, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1525/1791 [6:42:37<1:10:13, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1526/1791 [6:42:53<1:09:57, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1526/1791 [6:42:53<1:09:57, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1527/1791 [6:43:08<1:09:41, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1527/1791 [6:43:08<1:09:41, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1528/1791 [6:43:25<1:09:26, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1528/1791 [6:43:25<1:09:26, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1529/1791 [6:43:41<1:09:10, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1529/1791 [6:43:41<1:09:10, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1530/1791 [6:43:56<1:08:54, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1530/1791 [6:43:56<1:08:54, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1531/1791 [6:44:12<1:08:38, 0.06it/s, v_num=bc8f] Epoch 0: 85%|████████▌ | 1531/1791 [6:44:12<1:08:38, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1532/1791 [6:44:27<1:08:22, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1532/1791 [6:44:27<1:08:22, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1533/1791 [6:44:42<1:08:06, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1533/1791 [6:44:42<1:08:06, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1534/1791 [6:44:57<1:07:50, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1534/1791 [6:44:57<1:07:50, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1535/1791 [6:45:13<1:07:34, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1535/1791 [6:45:13<1:07:34, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1536/1791 [6:45:29<1:07:19, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1536/1791 [6:45:29<1:07:19, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1537/1791 [6:45:44<1:07:03, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1537/1791 [6:45:44<1:07:03, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1538/1791 [6:46:00<1:06:47, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1538/1791 [6:46:00<1:06:47, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1539/1791 [6:46:15<1:06:31, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1539/1791 [6:46:15<1:06:31, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1540/1791 [6:46:30<1:06:15, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1540/1791 [6:46:30<1:06:15, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1541/1791 [6:46:46<1:05:59, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1541/1791 [6:46:46<1:05:59, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1542/1791 [6:47:01<1:05:43, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1542/1791 [6:47:01<1:05:43, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1543/1791 [6:47:17<1:05:27, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1543/1791 [6:47:17<1:05:27, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1544/1791 [6:47:33<1:05:11, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▌ | 1544/1791 [6:47:33<1:05:11, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1545/1791 [6:47:48<1:04:55, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1545/1791 [6:47:48<1:04:55, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1546/1791 [6:48:03<1:04:40, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1546/1791 [6:48:03<1:04:40, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1547/1791 [6:48:19<1:04:24, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1547/1791 [6:48:19<1:04:24, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1548/1791 [6:48:34<1:04:08, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1548/1791 [6:48:34<1:04:08, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1549/1791 [6:48:49<1:03:52, 0.06it/s, v_num=bc8f] Epoch 0: 86%|████████▋ | 1549/1791 [6:48:49<1:03:52, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1550/1791 [6:49:04<1:03:36, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1550/1791 [6:49:04<1:03:36, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1551/1791 [6:49:20<1:03:20, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1551/1791 [6:49:20<1:03:20, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1552/1791 [6:49:36<1:03:04, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1552/1791 [6:49:36<1:03:04, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1553/1791 [6:49:52<1:02:48, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1553/1791 [6:49:52<1:02:48, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1554/1791 [6:50:07<1:02:32, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1554/1791 [6:50:07<1:02:32, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1555/1791 [6:50:22<1:02:16, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1555/1791 [6:50:22<1:02:16, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1556/1791 [6:50:38<1:02:01, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1556/1791 [6:50:38<1:02:01, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1557/1791 [6:50:53<1:01:45, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1557/1791 [6:50:53<1:01:45, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1558/1791 [6:51:08<1:01:29, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1558/1791 [6:51:08<1:01:29, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1559/1791 [6:51:23<1:01:13, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1559/1791 [6:51:23<1:01:13, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1560/1791 [6:51:40<1:00:57, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1560/1791 [6:51:40<1:00:57, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1561/1791 [6:51:56<1:00:41, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1561/1791 [6:51:56<1:00:41, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1562/1791 [6:52:11<1:00:25, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1562/1791 [6:52:11<1:00:25, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1563/1791 [6:52:26<1:00:09, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1563/1791 [6:52:26<1:00:09, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1564/1791 [6:52:41<59:53, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1564/1791 [6:52:41<59:53, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1565/1791 [6:52:56<59:37, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1565/1791 [6:52:56<59:37, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1566/1791 [6:53:11<59:22, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1566/1791 [6:53:11<59:22, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1567/1791 [6:53:26<59:06, 0.06it/s, v_num=bc8f] Epoch 0: 87%|████████▋ | 1567/1791 [6:53:26<59:06, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1568/1791 [6:53:43<58:50, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1568/1791 [6:53:43<58:50, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1569/1791 [6:53:58<58:34, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1569/1791 [6:53:58<58:34, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1570/1791 [6:54:13<58:18, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1570/1791 [6:54:13<58:18, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1571/1791 [6:54:29<58:02, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1571/1791 [6:54:29<58:02, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1572/1791 [6:54:45<57:46, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1572/1791 [6:54:45<57:46, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1573/1791 [6:55:00<57:30, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1573/1791 [6:55:00<57:30, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1574/1791 [6:55:16<57:15, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1574/1791 [6:55:16<57:15, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1575/1791 [6:55:31<56:59, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1575/1791 [6:55:31<56:59, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1576/1791 [6:55:47<56:43, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1576/1791 [6:55:47<56:43, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1577/1791 [6:56:02<56:27, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1577/1791 [6:56:02<56:27, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1578/1791 [6:56:17<56:11, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1578/1791 [6:56:17<56:11, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1579/1791 [6:56:33<55:55, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1579/1791 [6:56:33<55:55, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1580/1791 [6:56:48<55:39, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1580/1791 [6:56:48<55:39, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1581/1791 [6:57:05<55:24, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1581/1791 [6:57:05<55:24, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1582/1791 [6:57:19<55:08, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1582/1791 [6:57:19<55:08, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1583/1791 [6:57:35<54:52, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1583/1791 [6:57:35<54:52, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1584/1791 [6:57:52<54:36, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1584/1791 [6:57:52<54:36, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1585/1791 [6:58:07<54:20, 0.06it/s, v_num=bc8f] Epoch 0: 88%|████████▊ | 1585/1791 [6:58:07<54:20, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1586/1791 [6:58:22<54:04, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1586/1791 [6:58:22<54:04, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1587/1791 [6:58:37<53:48, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1587/1791 [6:58:37<53:48, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1588/1791 [6:58:53<53:32, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1588/1791 [6:58:53<53:32, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1589/1791 [6:59:08<53:16, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▊ | 1589/1791 [6:59:08<53:16, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1590/1791 [6:59:24<53:01, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1590/1791 [6:59:24<53:01, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1591/1791 [6:59:39<52:45, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1591/1791 [6:59:39<52:45, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1592/1791 [6:59:56<52:29, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1592/1791 [6:59:56<52:29, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1593/1791 [7:00:11<52:13, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1593/1791 [7:00:11<52:13, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1594/1791 [7:00:26<51:57, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1594/1791 [7:00:26<51:57, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1595/1791 [7:00:42<51:41, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1595/1791 [7:00:42<51:41, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1596/1791 [7:00:57<51:25, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1596/1791 [7:00:57<51:25, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1597/1791 [7:01:12<51:10, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1597/1791 [7:01:12<51:10, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1598/1791 [7:01:27<50:54, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1598/1791 [7:01:27<50:54, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1599/1791 [7:01:42<50:38, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1599/1791 [7:01:42<50:38, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1600/1791 [7:01:59<50:22, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1600/1791 [7:01:59<50:22, 0.06it/s, v_num=bc8f]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 + Epoch 0: 89%|████████▉ | 1601/1791 [7:02:31<50:08, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1601/1791 [7:02:31<50:08, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1602/1791 [7:02:47<49:52, 0.06it/s, v_num=bc8f] Epoch 0: 89%|████████▉ | 1602/1791 [7:02:47<49:52, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1603/1791 [7:03:03<49:37, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1603/1791 [7:03:03<49:37, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1604/1791 [7:03:19<49:21, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1604/1791 [7:03:19<49:21, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1605/1791 [7:03:35<49:05, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1605/1791 [7:03:35<49:05, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1606/1791 [7:03:51<48:49, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1606/1791 [7:03:51<48:49, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1607/1791 [7:04:07<48:33, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1607/1791 [7:04:07<48:33, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1608/1791 [7:04:25<48:18, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1608/1791 [7:04:25<48:18, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1609/1791 [7:04:40<48:02, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1609/1791 [7:04:40<48:02, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1610/1791 [7:04:56<47:46, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1610/1791 [7:04:56<47:46, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1611/1791 [7:05:11<47:30, 0.06it/s, v_num=bc8f] Epoch 0: 90%|████████▉ | 1611/1791 [7:05:11<47:30, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1612/1791 [7:05:27<47:14, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1612/1791 [7:05:27<47:14, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1613/1791 [7:05:42<46:58, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1613/1791 [7:05:42<46:58, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1614/1791 [7:05:57<46:42, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1614/1791 [7:05:57<46:42, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1615/1791 [7:06:12<46:26, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1615/1791 [7:06:12<46:26, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1616/1791 [7:06:29<46:11, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1616/1791 [7:06:29<46:11, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1617/1791 [7:06:45<45:55, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1617/1791 [7:06:45<45:55, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1618/1791 [7:07:00<45:39, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1618/1791 [7:07:00<45:39, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1619/1791 [7:07:15<45:23, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1619/1791 [7:07:15<45:23, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1620/1791 [7:07:31<45:07, 0.06it/s, v_num=bc8f] Epoch 0: 90%|█████████ | 1620/1791 [7:07:31<45:07, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1621/1791 [7:07:47<44:51, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1621/1791 [7:07:47<44:51, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1622/1791 [7:08:02<44:35, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1622/1791 [7:08:02<44:35, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1623/1791 [7:08:17<44:19, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1623/1791 [7:08:17<44:19, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1624/1791 [7:08:34<44:04, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1624/1791 [7:08:34<44:04, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1625/1791 [7:08:49<43:48, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1625/1791 [7:08:49<43:48, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1626/1791 [7:09:04<43:32, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1626/1791 [7:09:04<43:32, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1627/1791 [7:09:20<43:16, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1627/1791 [7:09:20<43:16, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1628/1791 [7:09:36<43:00, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1628/1791 [7:09:36<43:00, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1629/1791 [7:09:51<42:44, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1629/1791 [7:09:51<42:44, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1630/1791 [7:10:06<42:29, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1630/1791 [7:10:06<42:29, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1631/1791 [7:10:22<42:13, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1631/1791 [7:10:22<42:13, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1632/1791 [7:10:38<41:57, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1632/1791 [7:10:38<41:57, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1633/1791 [7:10:53<41:41, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1633/1791 [7:10:53<41:41, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1634/1791 [7:11:09<41:25, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████ | 1634/1791 [7:11:09<41:25, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1635/1791 [7:11:24<41:09, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1635/1791 [7:11:24<41:09, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1636/1791 [7:11:39<40:53, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1636/1791 [7:11:39<40:53, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1637/1791 [7:11:55<40:37, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1637/1791 [7:11:55<40:37, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1638/1791 [7:12:10<40:22, 0.06it/s, v_num=bc8f] Epoch 0: 91%|█████████▏| 1638/1791 [7:12:10<40:22, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1639/1791 [7:12:25<40:06, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1639/1791 [7:12:25<40:06, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1640/1791 [7:12:41<39:50, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1640/1791 [7:12:41<39:50, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1641/1791 [7:12:57<39:34, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1641/1791 [7:12:57<39:34, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1642/1791 [7:13:12<39:18, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1642/1791 [7:13:12<39:18, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1643/1791 [7:13:28<39:02, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1643/1791 [7:13:28<39:02, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1644/1791 [7:13:43<38:46, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1644/1791 [7:13:43<38:46, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1645/1791 [7:13:58<38:30, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1645/1791 [7:13:58<38:30, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1646/1791 [7:14:13<38:15, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1646/1791 [7:14:13<38:15, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1647/1791 [7:14:28<37:59, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1647/1791 [7:14:28<37:59, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1648/1791 [7:14:45<37:43, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1648/1791 [7:14:45<37:43, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1649/1791 [7:15:00<37:27, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1649/1791 [7:15:00<37:27, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1650/1791 [7:15:15<37:11, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1650/1791 [7:15:15<37:11, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1651/1791 [7:15:31<36:55, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1651/1791 [7:15:31<36:55, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1652/1791 [7:15:47<36:40, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1652/1791 [7:15:47<36:40, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1653/1791 [7:16:02<36:24, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1653/1791 [7:16:02<36:24, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1654/1791 [7:16:17<36:08, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1654/1791 [7:16:17<36:08, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1655/1791 [7:16:32<35:52, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1655/1791 [7:16:32<35:52, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1656/1791 [7:16:49<35:36, 0.06it/s, v_num=bc8f] Epoch 0: 92%|█████████▏| 1656/1791 [7:16:49<35:36, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1657/1791 [7:17:04<35:20, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1657/1791 [7:17:04<35:20, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1658/1791 [7:17:19<35:04, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1658/1791 [7:17:19<35:04, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1659/1791 [7:17:34<34:48, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1659/1791 [7:17:34<34:48, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1660/1791 [7:17:50<34:33, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1660/1791 [7:17:50<34:33, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1661/1791 [7:18:04<34:17, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1661/1791 [7:18:04<34:17, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1662/1791 [7:18:20<34:01, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1662/1791 [7:18:20<34:01, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1663/1791 [7:18:35<33:45, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1663/1791 [7:18:35<33:45, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1664/1791 [7:18:52<33:29, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1664/1791 [7:18:52<33:29, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1665/1791 [7:19:08<33:13, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1665/1791 [7:19:08<33:13, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1666/1791 [7:19:23<32:58, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1666/1791 [7:19:23<32:58, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1667/1791 [7:19:39<32:42, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1667/1791 [7:19:39<32:42, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1668/1791 [7:19:54<32:26, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1668/1791 [7:19:54<32:26, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1669/1791 [7:20:09<32:10, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1669/1791 [7:20:09<32:10, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1670/1791 [7:20:25<31:54, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1670/1791 [7:20:25<31:54, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1671/1791 [7:20:40<31:38, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1671/1791 [7:20:40<31:38, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1672/1791 [7:20:57<31:23, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1672/1791 [7:20:57<31:23, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1673/1791 [7:21:12<31:07, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1673/1791 [7:21:12<31:07, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1674/1791 [7:21:28<30:51, 0.06it/s, v_num=bc8f] Epoch 0: 93%|█████████▎| 1674/1791 [7:21:28<30:51, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1675/1791 [7:21:43<30:35, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1675/1791 [7:21:43<30:35, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1676/1791 [7:21:59<30:19, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1676/1791 [7:21:59<30:19, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1677/1791 [7:22:14<30:03, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1677/1791 [7:22:14<30:03, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1678/1791 [7:22:30<29:47, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1678/1791 [7:22:30<29:47, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1679/1791 [7:22:45<29:32, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▎| 1679/1791 [7:22:45<29:32, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1680/1791 [7:23:02<29:16, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1680/1791 [7:23:02<29:16, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1681/1791 [7:23:17<29:00, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1681/1791 [7:23:17<29:00, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1682/1791 [7:23:32<28:44, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1682/1791 [7:23:32<28:44, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1683/1791 [7:23:48<28:28, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1683/1791 [7:23:48<28:28, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1684/1791 [7:24:03<28:12, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1684/1791 [7:24:03<28:12, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1685/1791 [7:24:18<27:57, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1685/1791 [7:24:18<27:57, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1686/1791 [7:24:33<27:41, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1686/1791 [7:24:33<27:41, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1687/1791 [7:24:49<27:25, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1687/1791 [7:24:49<27:25, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1688/1791 [7:25:06<27:09, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1688/1791 [7:25:06<27:09, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1689/1791 [7:25:21<26:53, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1689/1791 [7:25:21<26:53, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1690/1791 [7:25:37<26:37, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1690/1791 [7:25:37<26:37, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1691/1791 [7:25:52<26:22, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1691/1791 [7:25:52<26:22, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1692/1791 [7:26:07<26:06, 0.06it/s, v_num=bc8f] Epoch 0: 94%|█████████▍| 1692/1791 [7:26:07<26:06, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1693/1791 [7:26:22<25:50, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1693/1791 [7:26:22<25:50, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1694/1791 [7:26:38<25:34, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1694/1791 [7:26:38<25:34, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1695/1791 [7:26:53<25:18, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1695/1791 [7:26:53<25:18, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1696/1791 [7:27:10<25:02, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1696/1791 [7:27:10<25:02, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1697/1791 [7:27:25<24:47, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1697/1791 [7:27:25<24:47, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1698/1791 [7:27:40<24:31, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1698/1791 [7:27:40<24:31, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1699/1791 [7:27:55<24:15, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1699/1791 [7:27:55<24:15, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1700/1791 [7:28:11<23:59, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1700/1791 [7:28:11<23:59, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1701/1791 [7:28:26<23:43, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▍| 1701/1791 [7:28:26<23:43, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1702/1791 [7:28:41<23:27, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1702/1791 [7:28:41<23:27, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1703/1791 [7:28:56<23:11, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1703/1791 [7:28:56<23:11, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1704/1791 [7:29:13<22:56, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1704/1791 [7:29:13<22:56, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1705/1791 [7:29:28<22:40, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1705/1791 [7:29:28<22:40, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1706/1791 [7:29:44<22:24, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1706/1791 [7:29:44<22:24, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1707/1791 [7:29:59<22:08, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1707/1791 [7:29:59<22:08, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1708/1791 [7:30:14<21:52, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1708/1791 [7:30:14<21:52, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1709/1791 [7:30:30<21:36, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1709/1791 [7:30:30<21:36, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1710/1791 [7:30:45<21:21, 0.06it/s, v_num=bc8f] Epoch 0: 95%|█████████▌| 1710/1791 [7:30:45<21:21, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1711/1791 [7:31:01<21:05, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1711/1791 [7:31:01<21:05, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1712/1791 [7:31:17<20:49, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1712/1791 [7:31:17<20:49, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1713/1791 [7:31:33<20:33, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1713/1791 [7:31:33<20:33, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1714/1791 [7:31:48<20:17, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1714/1791 [7:31:48<20:17, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1715/1791 [7:32:03<20:01, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1715/1791 [7:32:03<20:01, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1716/1791 [7:32:17<19:46, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1716/1791 [7:32:17<19:46, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1717/1791 [7:32:32<19:30, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1717/1791 [7:32:32<19:30, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1718/1791 [7:32:47<19:14, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1718/1791 [7:32:47<19:14, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1719/1791 [7:33:02<18:58, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1719/1791 [7:33:02<18:58, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1720/1791 [7:33:19<18:42, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1720/1791 [7:33:19<18:42, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1721/1791 [7:33:34<18:26, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1721/1791 [7:33:34<18:26, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1722/1791 [7:33:49<18:11, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1722/1791 [7:33:49<18:11, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1723/1791 [7:34:04<17:55, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▌| 1723/1791 [7:34:04<17:55, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1724/1791 [7:34:19<17:39, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1724/1791 [7:34:19<17:39, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1725/1791 [7:34:35<17:23, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1725/1791 [7:34:35<17:23, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1726/1791 [7:34:50<17:07, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1726/1791 [7:34:50<17:07, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1727/1791 [7:35:05<16:51, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1727/1791 [7:35:05<16:51, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1728/1791 [7:35:21<16:36, 0.06it/s, v_num=bc8f] Epoch 0: 96%|█████████▋| 1728/1791 [7:35:21<16:36, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1729/1791 [7:35:37<16:20, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1729/1791 [7:35:37<16:20, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1730/1791 [7:35:52<16:04, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1730/1791 [7:35:52<16:04, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1731/1791 [7:36:07<15:48, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1731/1791 [7:36:07<15:48, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1732/1791 [7:36:22<15:32, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1732/1791 [7:36:22<15:32, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1733/1791 [7:36:38<15:16, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1733/1791 [7:36:38<15:16, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1734/1791 [7:36:53<15:01, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1734/1791 [7:36:53<15:01, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1735/1791 [7:37:08<14:45, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1735/1791 [7:37:08<14:45, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1736/1791 [7:37:25<14:29, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1736/1791 [7:37:25<14:29, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1737/1791 [7:37:40<14:13, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1737/1791 [7:37:40<14:13, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1738/1791 [7:37:55<13:57, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1738/1791 [7:37:55<13:57, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1739/1791 [7:38:11<13:42, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1739/1791 [7:38:11<13:42, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1740/1791 [7:38:26<13:26, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1740/1791 [7:38:26<13:26, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1741/1791 [7:38:41<13:10, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1741/1791 [7:38:41<13:10, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1742/1791 [7:38:56<12:54, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1742/1791 [7:38:56<12:54, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1743/1791 [7:39:12<12:38, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1743/1791 [7:39:12<12:38, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1744/1791 [7:39:28<12:22, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1744/1791 [7:39:28<12:22, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1745/1791 [7:39:44<12:07, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1745/1791 [7:39:44<12:07, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1746/1791 [7:39:59<11:51, 0.06it/s, v_num=bc8f] Epoch 0: 97%|█████████▋| 1746/1791 [7:39:59<11:51, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1747/1791 [7:40:15<11:35, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1747/1791 [7:40:15<11:35, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1748/1791 [7:40:30<11:19, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1748/1791 [7:40:30<11:19, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1749/1791 [7:40:45<11:03, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1749/1791 [7:40:45<11:03, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1750/1791 [7:41:00<10:48, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1750/1791 [7:41:00<10:48, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1751/1791 [7:41:16<10:32, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1751/1791 [7:41:16<10:32, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1752/1791 [7:41:32<10:16, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1752/1791 [7:41:32<10:16, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1753/1791 [7:41:47<10:00, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1753/1791 [7:41:47<10:00, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1754/1791 [7:42:02<09:44, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1754/1791 [7:42:02<09:44, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1755/1791 [7:42:18<09:28, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1755/1791 [7:42:18<09:28, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1756/1791 [7:42:33<09:13, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1756/1791 [7:42:33<09:13, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1757/1791 [7:42:49<08:57, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1757/1791 [7:42:49<08:57, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1758/1791 [7:43:05<08:41, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1758/1791 [7:43:05<08:41, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1759/1791 [7:43:20<08:25, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1759/1791 [7:43:20<08:25, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1760/1791 [7:43:37<08:09, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1760/1791 [7:43:37<08:09, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1761/1791 [7:43:52<07:54, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1761/1791 [7:43:52<07:54, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1762/1791 [7:44:07<07:38, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1762/1791 [7:44:07<07:38, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1763/1791 [7:44:23<07:22, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1763/1791 [7:44:23<07:22, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1764/1791 [7:44:38<07:06, 0.06it/s, v_num=bc8f] Epoch 0: 98%|█████████▊| 1764/1791 [7:44:38<07:06, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1765/1791 [7:44:53<06:50, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1765/1791 [7:44:53<06:50, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1766/1791 [7:45:08<06:35, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1766/1791 [7:45:08<06:35, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1767/1791 [7:45:23<06:19, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1767/1791 [7:45:23<06:19, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1768/1791 [7:45:40<06:03, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▊| 1768/1791 [7:45:40<06:03, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1769/1791 [7:45:55<05:47, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1769/1791 [7:45:55<05:47, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1770/1791 [7:46:11<05:31, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1770/1791 [7:46:11<05:31, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1771/1791 [7:46:26<05:16, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1771/1791 [7:46:26<05:16, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1772/1791 [7:46:41<05:00, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1772/1791 [7:46:41<05:00, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1773/1791 [7:46:57<04:44, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1773/1791 [7:46:57<04:44, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1774/1791 [7:47:12<04:28, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1774/1791 [7:47:12<04:28, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1775/1791 [7:47:27<04:12, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1775/1791 [7:47:27<04:12, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1776/1791 [7:47:44<03:57, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1776/1791 [7:47:44<03:57, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1777/1791 [7:47:59<03:41, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1777/1791 [7:47:59<03:41, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1778/1791 [7:48:14<03:25, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1778/1791 [7:48:14<03:25, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1779/1791 [7:48:29<03:09, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1779/1791 [7:48:29<03:09, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1780/1791 [7:48:43<02:53, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1780/1791 [7:48:43<02:53, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1781/1791 [7:48:59<02:37, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1781/1791 [7:48:59<02:37, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1782/1791 [7:49:13<02:22, 0.06it/s, v_num=bc8f] Epoch 0: 99%|█████████▉| 1782/1791 [7:49:13<02:22, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1783/1791 [7:49:28<02:06, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1783/1791 [7:49:28<02:06, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1784/1791 [7:49:45<01:50, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1784/1791 [7:49:45<01:50, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1785/1791 [7:50:01<01:34, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1785/1791 [7:50:01<01:34, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1786/1791 [7:50:16<01:18, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1786/1791 [7:50:16<01:18, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1787/1791 [7:50:32<01:03, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1787/1791 [7:50:32<01:03, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1788/1791 [7:50:47<00:47, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1788/1791 [7:50:47<00:47, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1789/1791 [7:51:02<00:31, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1789/1791 [7:51:02<00:31, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1790/1791 [7:51:18<00:15, 0.06it/s, v_num=bc8f] Epoch 0: 100%|█████████▉| 1790/1791 [7:51:18<00:15, 0.06it/s, v_num=bc8f] Epoch 0: 100%|██████████| 1791/1791 [7:51:34<00:00, 0.06it/s, v_num=bc8f] Epoch 0: 100%|██████████| 1791/1791 [7:51:34<00:00, 0.06it/s, v_num=bc8f] Epoch 0: 100%|██████████| 1791/1791 [7:51:34<00:00, 0.06it/s, v_num=bc8f]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +`Trainer.fit` stopped: `max_epochs=1` reached. + Epoch 0: 100%|██████████| 1791/1791 [7:51:41<00:00, 0.06it/s, v_num=bc8f] +wandb: +wandb: 🚀 View run Test-AC-critic_expectile_0.9-inv_temp_1.0 at:  diff --git a/Llama3-8B-I_Word_log.txt b/Llama3-8B-I_Word_log.txt new file mode 100644 index 0000000000000000000000000000000000000000..c2fdcabadbb67765a1606d78f649a7cd3b64e838 --- /dev/null +++ b/Llama3-8B-I_Word_log.txt @@ -0,0 +1,168 @@ +The following values were not passed to `accelerate launch` and had defaults used instead: + `--num_cpu_threads_per_process` was set to `16` to improve out-of-box performance when training on CPUs +To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`. +[2025-09-15 17:09:21,177] [INFO] [real_accelerator.py:260:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-09-15 17:09:22,994] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +W0915 17:09:23.140565 1515135 site-packages/torch/distributed/run.py:774] +W0915 17:09:23.140565 1515135 site-packages/torch/distributed/run.py:774] ***************************************** +W0915 17:09:23.140565 1515135 site-packages/torch/distributed/run.py:774] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. +W0915 17:09:23.140565 1515135 site-packages/torch/distributed/run.py:774] ***************************************** +[rank: 2] Seed set to 42 +/home/jiashuo/codes/OfflineArcher/main.py:15: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead. + scaler = GradScaler(enabled=True) # 添加 AMP scaler(与 bf16 兼容) +[rank: 1] Seed set to 42 +[rank: 3] Seed set to 42 +/home/jiashuo/codes/OfflineArcher/main.py:15: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead. + scaler = GradScaler(enabled=True) # 添加 AMP scaler(与 bf16 兼容) +/home/jiashuo/codes/OfflineArcher/main.py:15: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead. + scaler = GradScaler(enabled=True) # 添加 AMP scaler(与 bf16 兼容) +[rank: 0] Seed set to 42 +/home/jiashuo/codes/OfflineArcher/main.py:15: FutureWarning: `torch.cuda.amp.GradScaler(args...)` is deprecated. Please use `torch.amp.GradScaler('cuda', args...)` instead. + scaler = GradScaler(enabled=True) # 添加 AMP scaler(与 bf16 兼容) +[rank: 2] Seed set to 42 +`torch_dtype` is deprecated! Use `dtype` instead! + Loading checkpoint shards: 0%| | 0/4 [00:00 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 0%| | 8/1801 [01:46<6:38:43, 0.07it/s, v_num=ufl5] Epoch 0: 0%| | 8/1801 [01:46<6:38:43, 0.07it/s, v_num=ufl5] Epoch 0: 0%| | 9/1801 [01:57<6:28:42, 0.08it/s, v_num=ufl5] Epoch 0: 0%| | 9/1801 [01:57<6:28:43, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 10/1801 [02:07<6:21:16, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 10/1801 [02:07<6:21:16, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 11/1801 [02:19<6:17:16, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 11/1801 [02:19<6:17:17, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 12/1801 [02:29<6:12:24, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 12/1801 [02:29<6:12:24, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 13/1801 [02:40<6:08:08, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 13/1801 [02:40<6:08:08, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 14/1801 [02:51<6:05:10, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 14/1801 [02:51<6:05:11, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 15/1801 [03:02<6:02:42, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 15/1801 [03:02<6:02:42, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 16/1801 [03:14<6:02:33, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 16/1801 [03:14<6:02:33, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 17/1801 [03:25<6:00:16, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 17/1801 [03:25<6:00:16, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 18/1801 [03:36<5:57:08, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 18/1801 [03:36<5:57:08, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 19/1801 [03:47<5:54:56, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 19/1801 [03:47<5:54:56, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 20/1801 [03:57<5:52:59, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 20/1801 [03:57<5:52:59, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 21/1801 [04:09<5:52:51, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 21/1801 [04:09<5:52:51, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 22/1801 [04:21<5:52:30, 0.08it/s, v_num=ufl5] Epoch 0: 1%| | 22/1801 [04:21<5:52:30, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 23/1801 [04:32<5:51:43, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 23/1801 [04:32<5:51:43, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 24/1801 [04:45<5:52:27, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 24/1801 [04:45<5:52:27, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 25/1801 [04:56<5:51:34, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 25/1801 [04:56<5:51:34, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 26/1801 [05:08<5:50:35, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 26/1801 [05:08<5:50:35, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 27/1801 [05:18<5:48:55, 0.08it/s, v_num=ufl5] Epoch 0: 1%|▏ | 27/1801 [05:18<5:48:55, 0.08it/s, v_num=ufl5] Epoch 0: 2%|▏ | 28/1801 [05:29<5:47:37, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 28/1801 [05:29<5:47:37, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 29/1801 [05:40<5:46:37, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 29/1801 [05:40<5:46:37, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 30/1801 [05:51<5:45:42, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 30/1801 [05:51<5:45:42, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 31/1801 [06:02<5:45:04, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 31/1801 [06:02<5:45:04, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 32/1801 [06:14<5:45:30, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 32/1801 [06:14<5:45:30, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 33/1801 [06:25<5:44:21, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 33/1801 [06:25<5:44:21, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 34/1801 [06:36<5:43:43, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 34/1801 [06:36<5:43:43, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 35/1801 [06:47<5:42:58, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 35/1801 [06:47<5:42:58, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 36/1801 [06:59<5:42:34, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 36/1801 [06:59<5:42:34, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 37/1801 [07:10<5:41:53, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 37/1801 [07:10<5:41:53, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 38/1801 [07:20<5:40:54, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 38/1801 [07:20<5:40:54, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 39/1801 [07:31<5:39:57, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 39/1801 [07:31<5:39:57, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 40/1801 [07:43<5:40:02, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 40/1801 [07:43<5:40:02, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 41/1801 [07:54<5:39:18, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 41/1801 [07:54<5:39:18, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 42/1801 [08:05<5:39:02, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 42/1801 [08:05<5:39:02, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 43/1801 [08:17<5:39:15, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 43/1801 [08:17<5:39:15, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 44/1801 [08:28<5:38:33, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 44/1801 [08:28<5:38:33, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 45/1801 [08:39<5:37:47, 0.09it/s, v_num=ufl5] Epoch 0: 2%|▏ | 45/1801 [08:39<5:37:47, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 46/1801 [08:50<5:37:33, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 46/1801 [08:50<5:37:33, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 47/1801 [09:02<5:37:15, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 47/1801 [09:02<5:37:15, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 48/1801 [09:14<5:37:15, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 48/1801 [09:14<5:37:15, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 49/1801 [09:24<5:36:29, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 49/1801 [09:24<5:36:29, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 50/1801 [09:36<5:36:19, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 50/1801 [09:36<5:36:19, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 51/1801 [09:47<5:36:09, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 51/1801 [09:47<5:36:09, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 52/1801 [09:58<5:35:32, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 52/1801 [09:58<5:35:32, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 53/1801 [10:09<5:34:49, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 53/1801 [10:09<5:34:49, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 54/1801 [10:19<5:34:08, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 54/1801 [10:19<5:34:08, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 55/1801 [10:30<5:33:40, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 55/1801 [10:30<5:33:40, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 56/1801 [10:43<5:34:01, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 56/1801 [10:43<5:34:01, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 57/1801 [10:53<5:33:25, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 57/1801 [10:53<5:33:25, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 58/1801 [11:04<5:32:53, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 58/1801 [11:04<5:32:53, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 59/1801 [11:15<5:32:36, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 59/1801 [11:15<5:32:36, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 60/1801 [11:27<5:32:31, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 60/1801 [11:27<5:32:31, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 61/1801 [11:38<5:32:07, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 61/1801 [11:38<5:32:07, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 62/1801 [11:49<5:31:53, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 62/1801 [11:49<5:31:53, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 63/1801 [12:01<5:31:44, 0.09it/s, v_num=ufl5] Epoch 0: 3%|▎ | 63/1801 [12:01<5:31:44, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 64/1801 [12:13<5:31:56, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 64/1801 [12:13<5:31:56, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 65/1801 [12:24<5:31:33, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 65/1801 [12:24<5:31:33, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 66/1801 [12:35<5:30:58, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 66/1801 [12:35<5:30:58, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 67/1801 [12:46<5:30:40, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▎ | 67/1801 [12:46<5:30:40, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 68/1801 [12:57<5:30:27, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 68/1801 [12:57<5:30:27, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 69/1801 [13:09<5:30:09, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 69/1801 [13:09<5:30:09, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 70/1801 [13:20<5:29:51, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 70/1801 [13:20<5:29:51, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 71/1801 [13:31<5:29:21, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 71/1801 [13:31<5:29:21, 0.09it/s, v_num=ufl5]Token indices sequence length is longer than the specified maximum sequence length for this model (1028 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 4%|▍ | 72/1801 [13:43<5:29:38, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 72/1801 [13:43<5:29:38, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 73/1801 [13:54<5:29:19, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 73/1801 [13:54<5:29:19, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 74/1801 [14:06<5:29:05, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 74/1801 [14:06<5:29:05, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 75/1801 [14:16<5:28:27, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 75/1801 [14:16<5:28:27, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 76/1801 [14:26<5:27:56, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 76/1801 [14:26<5:27:56, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 77/1801 [14:37<5:27:34, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 77/1801 [14:37<5:27:34, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 78/1801 [14:48<5:26:56, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 78/1801 [14:48<5:26:56, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 79/1801 [14:59<5:26:44, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 79/1801 [14:59<5:26:44, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 80/1801 [15:11<5:26:47, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 80/1801 [15:11<5:26:47, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 81/1801 [15:22<5:26:22, 0.09it/s, v_num=ufl5] Epoch 0: 4%|▍ | 81/1801 [15:22<5:26:22, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 82/1801 [15:32<5:25:52, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 82/1801 [15:32<5:25:52, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 83/1801 [15:43<5:25:30, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 83/1801 [15:43<5:25:30, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 84/1801 [15:54<5:25:19, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 84/1801 [15:54<5:25:19, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 85/1801 [16:06<5:25:16, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 85/1801 [16:06<5:25:16, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 86/1801 [16:17<5:24:55, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 86/1801 [16:17<5:24:55, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 87/1801 [16:28<5:24:35, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 87/1801 [16:28<5:24:35, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 88/1801 [16:41<5:24:57, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 88/1801 [16:41<5:24:57, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 89/1801 [16:52<5:24:43, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 89/1801 [16:52<5:24:43, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 90/1801 [17:04<5:24:40, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▍ | 90/1801 [17:04<5:24:40, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 91/1801 [17:16<5:24:40, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 91/1801 [17:16<5:24:40, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 92/1801 [17:27<5:24:26, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 92/1801 [17:27<5:24:26, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 93/1801 [17:39<5:24:17, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 93/1801 [17:39<5:24:17, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 94/1801 [17:50<5:24:00, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 94/1801 [17:50<5:24:00, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 95/1801 [18:01<5:23:44, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 95/1801 [18:01<5:23:44, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 96/1801 [18:14<5:23:54, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 96/1801 [18:14<5:23:54, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 97/1801 [18:25<5:23:47, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 97/1801 [18:25<5:23:47, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 98/1801 [18:37<5:23:31, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 98/1801 [18:37<5:23:31, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 99/1801 [18:47<5:23:06, 0.09it/s, v_num=ufl5] Epoch 0: 5%|▌ | 99/1801 [18:47<5:23:06, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 100/1801 [18:58<5:22:46, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 100/1801 [18:58<5:22:46, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 101/1801 [19:09<5:22:26, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 101/1801 [19:09<5:22:26, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 102/1801 [19:20<5:22:04, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 102/1801 [19:20<5:22:04, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 103/1801 [19:31<5:21:49, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 103/1801 [19:31<5:21:49, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 104/1801 [19:43<5:21:54, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 104/1801 [19:43<5:21:54, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 105/1801 [19:54<5:21:39, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 105/1801 [19:54<5:21:39, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 106/1801 [20:06<5:21:31, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 106/1801 [20:06<5:21:31, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 107/1801 [20:17<5:21:12, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 107/1801 [20:17<5:21:12, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 108/1801 [20:27<5:20:48, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 108/1801 [20:27<5:20:48, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 109/1801 [20:38<5:20:25, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 109/1801 [20:38<5:20:25, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 110/1801 [20:49<5:20:12, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 110/1801 [20:49<5:20:12, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 111/1801 [21:01<5:20:01, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 111/1801 [21:01<5:20:01, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 112/1801 [21:13<5:20:07, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▌ | 112/1801 [21:13<5:20:07, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 113/1801 [21:24<5:19:54, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 113/1801 [21:24<5:19:54, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 114/1801 [21:35<5:19:35, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 114/1801 [21:35<5:19:35, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 115/1801 [21:47<5:19:22, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 115/1801 [21:47<5:19:22, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 116/1801 [21:58<5:19:12, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 116/1801 [21:58<5:19:12, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 117/1801 [22:10<5:19:03, 0.09it/s, v_num=ufl5] Epoch 0: 6%|▋ | 117/1801 [22:10<5:19:03, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 118/1801 [22:20<5:18:44, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 118/1801 [22:20<5:18:44, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 119/1801 [22:32<5:18:33, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 119/1801 [22:32<5:18:33, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 120/1801 [22:44<5:18:29, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 120/1801 [22:44<5:18:29, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 121/1801 [22:54<5:18:08, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 121/1801 [22:54<5:18:08, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 122/1801 [23:05<5:17:53, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 122/1801 [23:05<5:17:53, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 123/1801 [23:16<5:17:27, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 123/1801 [23:16<5:17:27, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 124/1801 [23:27<5:17:15, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 124/1801 [23:27<5:17:15, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 125/1801 [23:38<5:16:54, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 125/1801 [23:38<5:16:54, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 126/1801 [23:49<5:16:45, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 126/1801 [23:49<5:16:45, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 127/1801 [24:00<5:16:29, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 127/1801 [24:00<5:16:29, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 128/1801 [24:12<5:16:21, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 128/1801 [24:12<5:16:21, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 129/1801 [24:22<5:16:01, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 129/1801 [24:22<5:16:01, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 130/1801 [24:34<5:15:54, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 130/1801 [24:34<5:15:54, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 131/1801 [24:45<5:15:37, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 131/1801 [24:45<5:15:37, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 132/1801 [24:56<5:15:19, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 132/1801 [24:56<5:15:19, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 133/1801 [25:07<5:15:05, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 133/1801 [25:07<5:15:05, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 134/1801 [25:18<5:14:45, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 134/1801 [25:18<5:14:45, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 135/1801 [25:28<5:14:26, 0.09it/s, v_num=ufl5] Epoch 0: 7%|▋ | 135/1801 [25:28<5:14:26, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 136/1801 [25:41<5:14:31, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 136/1801 [25:41<5:14:31, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 137/1801 [25:51<5:14:07, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 137/1801 [25:51<5:14:07, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 138/1801 [26:03<5:13:58, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 138/1801 [26:03<5:13:58, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 139/1801 [26:15<5:13:56, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 139/1801 [26:15<5:13:56, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 140/1801 [26:26<5:13:38, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 140/1801 [26:26<5:13:38, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 141/1801 [26:36<5:13:20, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 141/1801 [26:36<5:13:20, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 142/1801 [26:48<5:13:07, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 142/1801 [26:48<5:13:07, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 143/1801 [26:59<5:12:57, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 143/1801 [26:59<5:12:57, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 144/1801 [27:11<5:12:51, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 144/1801 [27:11<5:12:51, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 145/1801 [27:21<5:12:32, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 145/1801 [27:21<5:12:32, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 146/1801 [27:32<5:12:14, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 146/1801 [27:32<5:12:14, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 147/1801 [27:43<5:11:58, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 147/1801 [27:43<5:11:58, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 148/1801 [27:54<5:11:40, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 148/1801 [27:54<5:11:40, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 149/1801 [28:04<5:11:21, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 149/1801 [28:04<5:11:21, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 150/1801 [28:15<5:10:58, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 150/1801 [28:15<5:10:58, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 151/1801 [28:26<5:10:46, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 151/1801 [28:26<5:10:46, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 152/1801 [28:38<5:10:39, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 152/1801 [28:38<5:10:39, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 153/1801 [28:48<5:10:21, 0.09it/s, v_num=ufl5] Epoch 0: 8%|▊ | 153/1801 [28:48<5:10:21, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 154/1801 [29:00<5:10:09, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 154/1801 [29:00<5:10:09, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 155/1801 [29:11<5:09:57, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 155/1801 [29:11<5:09:57, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 156/1801 [29:22<5:09:45, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 156/1801 [29:22<5:09:45, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 157/1801 [29:33<5:09:30, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▊ | 157/1801 [29:33<5:09:30, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 158/1801 [29:45<5:09:22, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 158/1801 [29:45<5:09:22, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 159/1801 [29:55<5:09:02, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 159/1801 [29:55<5:09:02, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 160/1801 [30:07<5:08:59, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 160/1801 [30:07<5:08:59, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 161/1801 [30:18<5:08:44, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 161/1801 [30:18<5:08:44, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 162/1801 [30:29<5:08:33, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 162/1801 [30:29<5:08:33, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 163/1801 [30:40<5:08:16, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 163/1801 [30:40<5:08:16, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 164/1801 [30:52<5:08:07, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 164/1801 [30:52<5:08:07, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 165/1801 [31:02<5:07:47, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 165/1801 [31:02<5:07:47, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 166/1801 [31:14<5:07:38, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 166/1801 [31:14<5:07:38, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 167/1801 [31:25<5:07:24, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 167/1801 [31:25<5:07:24, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 168/1801 [31:37<5:07:23, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 168/1801 [31:37<5:07:23, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 169/1801 [31:48<5:07:10, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 169/1801 [31:48<5:07:10, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 170/1801 [32:00<5:07:01, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 170/1801 [32:00<5:07:01, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 171/1801 [32:11<5:06:55, 0.09it/s, v_num=ufl5] Epoch 0: 9%|▉ | 171/1801 [32:11<5:06:55, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 172/1801 [32:22<5:06:38, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 172/1801 [32:22<5:06:38, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 173/1801 [32:33<5:06:24, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 173/1801 [32:33<5:06:24, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 174/1801 [32:44<5:06:13, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 174/1801 [32:44<5:06:13, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 175/1801 [32:55<5:05:55, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 175/1801 [32:55<5:05:55, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 176/1801 [33:08<5:05:56, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 176/1801 [33:08<5:05:56, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 177/1801 [33:18<5:05:38, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 177/1801 [33:18<5:05:38, 0.09it/s, v_num=ufl5]Token indices sequence length is longer than the specified maximum sequence length for this model (1027 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 10%|▉ | 178/1801 [33:30<5:05:33, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 178/1801 [33:30<5:05:33, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 179/1801 [33:41<5:05:21, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 179/1801 [33:41<5:05:21, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 180/1801 [33:53<5:05:09, 0.09it/s, v_num=ufl5] Epoch 0: 10%|▉ | 180/1801 [33:53<5:05:09, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 181/1801 [34:03<5:04:50, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 181/1801 [34:03<5:04:50, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 182/1801 [34:15<5:04:41, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 182/1801 [34:15<5:04:41, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 183/1801 [34:25<5:04:24, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 183/1801 [34:25<5:04:24, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 184/1801 [34:37<5:04:20, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 184/1801 [34:37<5:04:20, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 185/1801 [34:48<5:04:02, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 185/1801 [34:48<5:04:02, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 186/1801 [34:59<5:03:50, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 186/1801 [34:59<5:03:50, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 187/1801 [35:10<5:03:33, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 187/1801 [35:10<5:03:33, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 188/1801 [35:21<5:03:20, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 188/1801 [35:21<5:03:20, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 189/1801 [35:33<5:03:18, 0.09it/s, v_num=ufl5] Epoch 0: 10%|█ | 189/1801 [35:33<5:03:18, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 190/1801 [35:44<5:03:01, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 190/1801 [35:44<5:03:01, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 191/1801 [35:55<5:02:50, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 191/1801 [35:55<5:02:50, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 192/1801 [36:07<5:02:47, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 192/1801 [36:07<5:02:47, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 193/1801 [36:18<5:02:32, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 193/1801 [36:18<5:02:32, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 194/1801 [36:29<5:02:15, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 194/1801 [36:29<5:02:15, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 195/1801 [36:40<5:02:03, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 195/1801 [36:40<5:02:03, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 196/1801 [36:51<5:01:51, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 196/1801 [36:51<5:01:51, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 197/1801 [37:03<5:01:42, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 197/1801 [37:03<5:01:42, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 198/1801 [37:14<5:01:26, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 198/1801 [37:14<5:01:26, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 199/1801 [37:24<5:01:07, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 199/1801 [37:24<5:01:07, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 200/1801 [37:36<5:01:03, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 200/1801 [37:36<5:01:03, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 201/1801 [37:47<5:00:49, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 201/1801 [37:47<5:00:49, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 202/1801 [37:58<5:00:33, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█ | 202/1801 [37:58<5:00:33, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 203/1801 [38:09<5:00:24, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 203/1801 [38:09<5:00:24, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 204/1801 [38:19<5:00:03, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 204/1801 [38:19<5:00:03, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 205/1801 [38:30<4:59:46, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 205/1801 [38:30<4:59:46, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 206/1801 [38:41<4:59:32, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 206/1801 [38:41<4:59:32, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 207/1801 [38:52<4:59:17, 0.09it/s, v_num=ufl5] Epoch 0: 11%|█▏ | 207/1801 [38:52<4:59:17, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 208/1801 [39:03<4:59:10, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 208/1801 [39:03<4:59:10, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 209/1801 [39:14<4:58:57, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 209/1801 [39:14<4:58:57, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 210/1801 [39:26<4:58:46, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 210/1801 [39:26<4:58:46, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 211/1801 [39:36<4:58:27, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 211/1801 [39:36<4:58:27, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 212/1801 [39:47<4:58:16, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 212/1801 [39:47<4:58:16, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 213/1801 [39:59<4:58:07, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 213/1801 [39:59<4:58:07, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 214/1801 [40:10<4:57:55, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 214/1801 [40:10<4:57:55, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 215/1801 [40:21<4:57:41, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 215/1801 [40:21<4:57:41, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 216/1801 [40:33<4:57:36, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 216/1801 [40:33<4:57:36, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 217/1801 [40:44<4:57:21, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 217/1801 [40:44<4:57:21, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 218/1801 [40:54<4:57:06, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 218/1801 [40:54<4:57:06, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 219/1801 [41:05<4:56:49, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 219/1801 [41:05<4:56:49, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 220/1801 [41:16<4:56:39, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 220/1801 [41:16<4:56:39, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 221/1801 [41:27<4:56:25, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 221/1801 [41:27<4:56:25, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 222/1801 [41:39<4:56:14, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 222/1801 [41:39<4:56:14, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 223/1801 [41:49<4:55:57, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 223/1801 [41:49<4:55:57, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 224/1801 [42:01<4:55:54, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 224/1801 [42:01<4:55:54, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 225/1801 [42:12<4:55:41, 0.09it/s, v_num=ufl5] Epoch 0: 12%|█▏ | 225/1801 [42:12<4:55:41, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 226/1801 [42:24<4:55:29, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 226/1801 [42:24<4:55:29, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 227/1801 [42:34<4:55:12, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 227/1801 [42:34<4:55:12, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 228/1801 [42:45<4:54:59, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 228/1801 [42:45<4:54:59, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 229/1801 [42:56<4:54:48, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 229/1801 [42:56<4:54:48, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 230/1801 [43:07<4:54:37, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 230/1801 [43:07<4:54:37, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 231/1801 [43:18<4:54:19, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 231/1801 [43:18<4:54:20, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 232/1801 [43:30<4:54:13, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 232/1801 [43:30<4:54:13, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 233/1801 [43:42<4:54:06, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 233/1801 [43:42<4:54:06, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 234/1801 [43:53<4:53:53, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 234/1801 [43:53<4:53:53, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 235/1801 [44:05<4:53:46, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 235/1801 [44:05<4:53:46, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 236/1801 [44:15<4:53:31, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 236/1801 [44:15<4:53:31, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 237/1801 [44:26<4:53:16, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 237/1801 [44:26<4:53:16, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 238/1801 [44:37<4:53:02, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 238/1801 [44:37<4:53:02, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 239/1801 [44:48<4:52:50, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 239/1801 [44:48<4:52:50, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 240/1801 [45:00<4:52:42, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 240/1801 [45:00<4:52:42, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 241/1801 [45:11<4:52:29, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 241/1801 [45:11<4:52:29, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 242/1801 [45:22<4:52:17, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 242/1801 [45:22<4:52:17, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 243/1801 [45:33<4:52:07, 0.09it/s, v_num=ufl5] Epoch 0: 13%|█▎ | 243/1801 [45:33<4:52:07, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 244/1801 [45:44<4:51:52, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 244/1801 [45:44<4:51:52, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 245/1801 [45:56<4:51:43, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 245/1801 [45:56<4:51:43, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 246/1801 [46:07<4:51:31, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 246/1801 [46:07<4:51:31, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 247/1801 [46:18<4:51:18, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▎ | 247/1801 [46:18<4:51:18, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 248/1801 [46:30<4:51:15, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 248/1801 [46:30<4:51:15, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 249/1801 [46:41<4:51:04, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 249/1801 [46:41<4:51:04, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 250/1801 [46:53<4:50:52, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 250/1801 [46:53<4:50:52, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 251/1801 [47:04<4:50:39, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 251/1801 [47:04<4:50:39, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 252/1801 [47:15<4:50:30, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 252/1801 [47:15<4:50:30, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 253/1801 [47:26<4:50:15, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 253/1801 [47:26<4:50:15, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 254/1801 [47:37<4:50:06, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 254/1801 [47:37<4:50:06, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 255/1801 [47:49<4:49:54, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 255/1801 [47:49<4:49:54, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 256/1801 [48:00<4:49:46, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 256/1801 [48:00<4:49:46, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 257/1801 [48:11<4:49:33, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 257/1801 [48:11<4:49:33, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 258/1801 [48:23<4:49:23, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 258/1801 [48:23<4:49:23, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 259/1801 [48:34<4:49:10, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 259/1801 [48:34<4:49:10, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 260/1801 [48:45<4:48:59, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 260/1801 [48:45<4:48:59, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 261/1801 [48:57<4:48:49, 0.09it/s, v_num=ufl5] Epoch 0: 14%|█▍ | 261/1801 [48:57<4:48:49, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 262/1801 [49:07<4:48:34, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 262/1801 [49:07<4:48:34, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 263/1801 [49:18<4:48:22, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 263/1801 [49:18<4:48:22, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 264/1801 [49:30<4:48:14, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 264/1801 [49:30<4:48:14, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 265/1801 [49:41<4:48:00, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 265/1801 [49:41<4:48:00, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 266/1801 [49:52<4:47:50, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 266/1801 [49:52<4:47:50, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 267/1801 [50:03<4:47:36, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 267/1801 [50:03<4:47:36, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 268/1801 [50:14<4:47:23, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 268/1801 [50:14<4:47:23, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 269/1801 [50:26<4:47:14, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 269/1801 [50:26<4:47:14, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 270/1801 [50:37<4:47:03, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▍ | 270/1801 [50:37<4:47:03, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 271/1801 [50:48<4:46:48, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 271/1801 [50:48<4:46:48, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 272/1801 [50:59<4:46:39, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 272/1801 [50:59<4:46:39, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 273/1801 [51:10<4:46:26, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 273/1801 [51:10<4:46:26, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 274/1801 [51:20<4:46:10, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 274/1801 [51:20<4:46:10, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 275/1801 [51:32<4:45:58, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 275/1801 [51:32<4:45:58, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 276/1801 [51:42<4:45:42, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 276/1801 [51:42<4:45:42, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 277/1801 [51:53<4:45:31, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 277/1801 [51:53<4:45:31, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 278/1801 [52:04<4:45:16, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 278/1801 [52:04<4:45:16, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 279/1801 [52:15<4:45:02, 0.09it/s, v_num=ufl5] Epoch 0: 15%|█▌ | 279/1801 [52:15<4:45:02, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 280/1801 [52:27<4:44:56, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 280/1801 [52:27<4:44:56, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 281/1801 [52:38<4:44:46, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 281/1801 [52:38<4:44:46, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 282/1801 [52:50<4:44:36, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 282/1801 [52:50<4:44:36, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 283/1801 [53:01<4:44:23, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 283/1801 [53:01<4:44:23, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 284/1801 [53:12<4:44:13, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 284/1801 [53:12<4:44:13, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 285/1801 [53:23<4:44:01, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 285/1801 [53:23<4:44:01, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 286/1801 [53:34<4:43:47, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 286/1801 [53:34<4:43:47, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 287/1801 [53:45<4:43:36, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 287/1801 [53:45<4:43:36, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 288/1801 [53:58<4:43:31, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 288/1801 [53:58<4:43:31, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 289/1801 [54:08<4:43:17, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 289/1801 [54:08<4:43:17, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 290/1801 [54:19<4:43:03, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 290/1801 [54:19<4:43:03, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 291/1801 [54:30<4:42:51, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 291/1801 [54:30<4:42:51, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 292/1801 [54:41<4:42:36, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▌ | 292/1801 [54:41<4:42:36, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 293/1801 [54:52<4:42:23, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 293/1801 [54:52<4:42:23, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 294/1801 [55:03<4:42:14, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 294/1801 [55:03<4:42:14, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 295/1801 [55:14<4:42:00, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 295/1801 [55:14<4:42:00, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 296/1801 [55:26<4:41:51, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 296/1801 [55:26<4:41:51, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 297/1801 [55:37<4:41:40, 0.09it/s, v_num=ufl5] Epoch 0: 16%|█▋ | 297/1801 [55:37<4:41:40, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 298/1801 [55:47<4:41:25, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 298/1801 [55:47<4:41:25, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 299/1801 [55:59<4:41:14, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 299/1801 [55:59<4:41:14, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 300/1801 [56:09<4:41:00, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 300/1801 [56:09<4:41:00, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 301/1801 [56:20<4:40:47, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 301/1801 [56:20<4:40:47, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 302/1801 [56:30<4:40:31, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 302/1801 [56:30<4:40:31, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 303/1801 [56:42<4:40:20, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 303/1801 [56:42<4:40:20, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 304/1801 [56:54<4:40:12, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 304/1801 [56:54<4:40:12, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 305/1801 [57:05<4:39:59, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 305/1801 [57:05<4:39:59, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 306/1801 [57:16<4:39:47, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 306/1801 [57:16<4:39:47, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 307/1801 [57:27<4:39:34, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 307/1801 [57:27<4:39:34, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 308/1801 [57:37<4:39:21, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 308/1801 [57:37<4:39:21, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 309/1801 [57:48<4:39:06, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 309/1801 [57:48<4:39:07, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 310/1801 [57:59<4:38:53, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 310/1801 [57:59<4:38:53, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 311/1801 [58:09<4:38:39, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 311/1801 [58:09<4:38:39, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 312/1801 [58:21<4:38:32, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 312/1801 [58:21<4:38:32, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 313/1801 [58:32<4:38:16, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 313/1801 [58:32<4:38:16, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 314/1801 [58:42<4:38:01, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 314/1801 [58:42<4:38:01, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 315/1801 [58:53<4:37:47, 0.09it/s, v_num=ufl5] Epoch 0: 17%|█▋ | 315/1801 [58:53<4:37:47, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 316/1801 [59:04<4:37:35, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 316/1801 [59:04<4:37:35, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 317/1801 [59:14<4:37:20, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 317/1801 [59:14<4:37:20, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 318/1801 [59:25<4:37:06, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 318/1801 [59:25<4:37:06, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 319/1801 [59:35<4:36:51, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 319/1801 [59:35<4:36:51, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 320/1801 [59:48<4:36:47, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 320/1801 [59:48<4:36:47, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 321/1801 [59:59<4:36:36, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 321/1801 [59:59<4:36:36, 0.09it/s, v_num=ufl5]Token indices sequence length is longer than the specified maximum sequence length for this model (1050 > 1024). Running this sequence through the model will result in indexing errors + Epoch 0: 18%|█▊ | 322/1801 [1:00:11<4:36:27, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 322/1801 [1:00:11<4:36:27, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 323/1801 [1:00:22<4:36:16, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 323/1801 [1:00:22<4:36:16, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 324/1801 [1:00:33<4:36:05, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 324/1801 [1:00:33<4:36:05, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 325/1801 [1:00:44<4:35:53, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 325/1801 [1:00:44<4:35:53, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 326/1801 [1:00:55<4:35:39, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 326/1801 [1:00:55<4:35:39, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 327/1801 [1:01:06<4:35:29, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 327/1801 [1:01:06<4:35:29, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 328/1801 [1:01:19<4:35:23, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 328/1801 [1:01:19<4:35:23, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 329/1801 [1:01:31<4:35:14, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 329/1801 [1:01:31<4:35:14, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 330/1801 [1:01:42<4:35:02, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 330/1801 [1:01:42<4:35:02, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 331/1801 [1:01:53<4:34:50, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 331/1801 [1:01:53<4:34:50, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 332/1801 [1:02:04<4:34:39, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 332/1801 [1:02:04<4:34:39, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 333/1801 [1:02:14<4:34:25, 0.09it/s, v_num=ufl5] Epoch 0: 18%|█▊ | 333/1801 [1:02:14<4:34:25, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 334/1801 [1:02:25<4:34:11, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 334/1801 [1:02:25<4:34:11, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 335/1801 [1:02:36<4:34:00, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 335/1801 [1:02:36<4:34:00, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 336/1801 [1:02:49<4:33:54, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 336/1801 [1:02:49<4:33:54, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 337/1801 [1:02:59<4:33:38, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▊ | 337/1801 [1:02:59<4:33:38, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 338/1801 [1:03:10<4:33:26, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 338/1801 [1:03:10<4:33:26, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 339/1801 [1:03:21<4:33:13, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 339/1801 [1:03:21<4:33:13, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 340/1801 [1:03:32<4:33:03, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 340/1801 [1:03:32<4:33:03, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 341/1801 [1:03:43<4:32:50, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 341/1801 [1:03:43<4:32:50, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 342/1801 [1:03:54<4:32:37, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 342/1801 [1:03:54<4:32:37, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 343/1801 [1:04:05<4:32:25, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 343/1801 [1:04:05<4:32:25, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 344/1801 [1:04:17<4:32:16, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 344/1801 [1:04:17<4:32:16, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 345/1801 [1:04:28<4:32:04, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 345/1801 [1:04:28<4:32:04, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 346/1801 [1:04:39<4:31:52, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 346/1801 [1:04:39<4:31:52, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 347/1801 [1:04:50<4:31:41, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 347/1801 [1:04:50<4:31:41, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 348/1801 [1:05:01<4:31:29, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 348/1801 [1:05:01<4:31:29, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 349/1801 [1:05:11<4:31:15, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 349/1801 [1:05:11<4:31:15, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 350/1801 [1:05:22<4:31:02, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 350/1801 [1:05:22<4:31:02, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 351/1801 [1:05:33<4:30:48, 0.09it/s, v_num=ufl5] Epoch 0: 19%|█▉ | 351/1801 [1:05:33<4:30:48, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 352/1801 [1:05:45<4:30:42, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 352/1801 [1:05:45<4:30:42, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 353/1801 [1:05:56<4:30:29, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 353/1801 [1:05:56<4:30:29, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 354/1801 [1:06:07<4:30:18, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 354/1801 [1:06:07<4:30:18, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 355/1801 [1:06:19<4:30:08, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 355/1801 [1:06:19<4:30:08, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 356/1801 [1:06:30<4:29:55, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 356/1801 [1:06:30<4:29:55, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 357/1801 [1:06:40<4:29:42, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 357/1801 [1:06:40<4:29:42, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 358/1801 [1:06:51<4:29:28, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 358/1801 [1:06:51<4:29:28, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 359/1801 [1:07:02<4:29:15, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 359/1801 [1:07:02<4:29:15, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 360/1801 [1:07:13<4:29:05, 0.09it/s, v_num=ufl5] Epoch 0: 20%|█▉ | 360/1801 [1:07:13<4:29:05, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 361/1801 [1:07:23<4:28:51, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 361/1801 [1:07:23<4:28:51, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 362/1801 [1:07:35<4:28:39, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 362/1801 [1:07:35<4:28:39, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 363/1801 [1:07:45<4:28:26, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 363/1801 [1:07:45<4:28:26, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 364/1801 [1:07:57<4:28:16, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 364/1801 [1:07:57<4:28:16, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 365/1801 [1:08:08<4:28:05, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 365/1801 [1:08:08<4:28:05, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 366/1801 [1:08:20<4:27:56, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 366/1801 [1:08:20<4:27:56, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 367/1801 [1:08:31<4:27:45, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 367/1801 [1:08:31<4:27:45, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 368/1801 [1:08:42<4:27:34, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 368/1801 [1:08:42<4:27:34, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 369/1801 [1:08:53<4:27:22, 0.09it/s, v_num=ufl5] Epoch 0: 20%|██ | 369/1801 [1:08:53<4:27:22, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 370/1801 [1:09:04<4:27:07, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 370/1801 [1:09:04<4:27:07, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 371/1801 [1:09:14<4:26:53, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 371/1801 [1:09:14<4:26:53, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 372/1801 [1:09:25<4:26:40, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 372/1801 [1:09:25<4:26:40, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 373/1801 [1:09:36<4:26:28, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 373/1801 [1:09:36<4:26:28, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 374/1801 [1:09:46<4:26:14, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 374/1801 [1:09:46<4:26:14, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 375/1801 [1:09:57<4:26:00, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 375/1801 [1:09:57<4:26:00, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 376/1801 [1:10:08<4:25:50, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 376/1801 [1:10:08<4:25:50, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 377/1801 [1:10:19<4:25:38, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 377/1801 [1:10:19<4:25:38, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 378/1801 [1:10:31<4:25:30, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 378/1801 [1:10:31<4:25:30, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 379/1801 [1:10:42<4:25:16, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 379/1801 [1:10:42<4:25:16, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 380/1801 [1:10:53<4:25:05, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 380/1801 [1:10:53<4:25:05, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 381/1801 [1:11:03<4:24:51, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 381/1801 [1:11:03<4:24:51, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 382/1801 [1:11:15<4:24:41, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██ | 382/1801 [1:11:15<4:24:41, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 383/1801 [1:11:25<4:24:27, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 383/1801 [1:11:25<4:24:27, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 384/1801 [1:11:38<4:24:20, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 384/1801 [1:11:38<4:24:20, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 385/1801 [1:11:48<4:24:07, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 385/1801 [1:11:48<4:24:07, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 386/1801 [1:11:59<4:23:53, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 386/1801 [1:11:59<4:23:53, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 387/1801 [1:12:09<4:23:39, 0.09it/s, v_num=ufl5] Epoch 0: 21%|██▏ | 387/1801 [1:12:09<4:23:39, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 388/1801 [1:12:20<4:23:26, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 388/1801 [1:12:20<4:23:26, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 389/1801 [1:12:31<4:23:15, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 389/1801 [1:12:31<4:23:15, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 390/1801 [1:12:42<4:23:04, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 390/1801 [1:12:42<4:23:04, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 391/1801 [1:12:52<4:22:49, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 391/1801 [1:12:52<4:22:49, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 392/1801 [1:13:04<4:22:40, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 392/1801 [1:13:04<4:22:40, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 393/1801 [1:13:15<4:22:26, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 393/1801 [1:13:15<4:22:26, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 394/1801 [1:13:25<4:22:12, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 394/1801 [1:13:25<4:22:12, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 395/1801 [1:13:35<4:21:57, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 395/1801 [1:13:35<4:21:57, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 396/1801 [1:13:47<4:21:47, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 396/1801 [1:13:47<4:21:47, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 397/1801 [1:13:57<4:21:34, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 397/1801 [1:13:57<4:21:34, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 398/1801 [1:14:08<4:21:21, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 398/1801 [1:14:08<4:21:21, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 399/1801 [1:14:19<4:21:09, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 399/1801 [1:14:19<4:21:09, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 400/1801 [1:14:31<4:21:00, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 400/1801 [1:14:31<4:21:00, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 401/1801 [1:14:42<4:20:49, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 401/1801 [1:14:42<4:20:49, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 402/1801 [1:14:53<4:20:36, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 402/1801 [1:14:53<4:20:36, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 403/1801 [1:15:03<4:20:23, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 403/1801 [1:15:03<4:20:23, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 404/1801 [1:15:14<4:20:09, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 404/1801 [1:15:14<4:20:09, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 405/1801 [1:15:24<4:19:56, 0.09it/s, v_num=ufl5] Epoch 0: 22%|██▏ | 405/1801 [1:15:24<4:19:56, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 406/1801 [1:15:36<4:19:45, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 406/1801 [1:15:36<4:19:45, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 407/1801 [1:15:47<4:19:36, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 407/1801 [1:15:47<4:19:36, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 408/1801 [1:16:00<4:19:29, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 408/1801 [1:16:00<4:19:29, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 409/1801 [1:16:11<4:19:17, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 409/1801 [1:16:11<4:19:17, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 410/1801 [1:16:22<4:19:07, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 410/1801 [1:16:22<4:19:07, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 411/1801 [1:16:33<4:18:56, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 411/1801 [1:16:33<4:18:56, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 412/1801 [1:16:44<4:18:44, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 412/1801 [1:16:44<4:18:44, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 413/1801 [1:16:56<4:18:35, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 413/1801 [1:16:56<4:18:35, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 414/1801 [1:17:07<4:18:21, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 414/1801 [1:17:07<4:18:21, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 415/1801 [1:17:18<4:18:09, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 415/1801 [1:17:18<4:18:09, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 416/1801 [1:17:29<4:18:00, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 416/1801 [1:17:29<4:18:00, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 417/1801 [1:17:39<4:17:46, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 417/1801 [1:17:39<4:17:46, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 418/1801 [1:17:51<4:17:34, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 418/1801 [1:17:51<4:17:34, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 419/1801 [1:18:01<4:17:22, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 419/1801 [1:18:01<4:17:22, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 420/1801 [1:18:13<4:17:11, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 420/1801 [1:18:13<4:17:11, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 421/1801 [1:18:24<4:17:00, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 421/1801 [1:18:24<4:17:00, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 422/1801 [1:18:35<4:16:49, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 422/1801 [1:18:35<4:16:49, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 423/1801 [1:18:45<4:16:35, 0.09it/s, v_num=ufl5] Epoch 0: 23%|██▎ | 423/1801 [1:18:45<4:16:35, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 424/1801 [1:18:57<4:16:25, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 424/1801 [1:18:57<4:16:25, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 425/1801 [1:19:08<4:16:13, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 425/1801 [1:19:08<4:16:13, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 426/1801 [1:19:18<4:15:59, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 426/1801 [1:19:18<4:15:59, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 427/1801 [1:19:29<4:15:47, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▎ | 427/1801 [1:19:29<4:15:47, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 428/1801 [1:19:40<4:15:37, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 428/1801 [1:19:40<4:15:37, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 429/1801 [1:19:52<4:15:25, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 429/1801 [1:19:52<4:15:25, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 430/1801 [1:20:02<4:15:13, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 430/1801 [1:20:02<4:15:13, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 431/1801 [1:20:14<4:15:03, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 431/1801 [1:20:14<4:15:03, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 432/1801 [1:20:26<4:14:55, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 432/1801 [1:20:26<4:14:55, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 433/1801 [1:20:38<4:14:46, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 433/1801 [1:20:38<4:14:46, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 434/1801 [1:20:49<4:14:34, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 434/1801 [1:20:49<4:14:34, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 435/1801 [1:21:00<4:14:22, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 435/1801 [1:21:00<4:14:22, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 436/1801 [1:21:10<4:14:07, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 436/1801 [1:21:10<4:14:07, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 437/1801 [1:21:20<4:13:54, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 437/1801 [1:21:20<4:13:54, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 438/1801 [1:21:32<4:13:45, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 438/1801 [1:21:32<4:13:45, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 439/1801 [1:21:43<4:13:33, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 439/1801 [1:21:43<4:13:33, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 440/1801 [1:21:56<4:13:26, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 440/1801 [1:21:56<4:13:26, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 441/1801 [1:22:06<4:13:13, 0.09it/s, v_num=ufl5] Epoch 0: 24%|██▍ | 441/1801 [1:22:06<4:13:13, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 442/1801 [1:22:18<4:13:02, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 442/1801 [1:22:18<4:13:02, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 443/1801 [1:22:28<4:12:50, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 443/1801 [1:22:28<4:12:50, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 444/1801 [1:22:39<4:12:37, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 444/1801 [1:22:39<4:12:37, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 445/1801 [1:22:50<4:12:24, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 445/1801 [1:22:50<4:12:24, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 446/1801 [1:23:00<4:12:11, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 446/1801 [1:23:00<4:12:11, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 447/1801 [1:23:11<4:11:58, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 447/1801 [1:23:11<4:11:58, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 448/1801 [1:23:23<4:11:50, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 448/1801 [1:23:23<4:11:50, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 449/1801 [1:23:33<4:11:37, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 449/1801 [1:23:33<4:11:37, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 450/1801 [1:23:44<4:11:23, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▍ | 450/1801 [1:23:44<4:11:23, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 451/1801 [1:23:55<4:11:12, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 451/1801 [1:23:55<4:11:12, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 452/1801 [1:24:06<4:11:01, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 452/1801 [1:24:06<4:11:01, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 453/1801 [1:24:17<4:10:49, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 453/1801 [1:24:17<4:10:49, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 454/1801 [1:24:27<4:10:35, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 454/1801 [1:24:27<4:10:35, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 455/1801 [1:24:38<4:10:24, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 455/1801 [1:24:38<4:10:24, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 456/1801 [1:24:50<4:10:15, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 456/1801 [1:24:50<4:10:15, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 457/1801 [1:25:01<4:10:01, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 457/1801 [1:25:01<4:10:01, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 458/1801 [1:25:12<4:09:50, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 458/1801 [1:25:12<4:09:50, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 459/1801 [1:25:22<4:09:36, 0.09it/s, v_num=ufl5] Epoch 0: 25%|██▌ | 459/1801 [1:25:22<4:09:36, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 460/1801 [1:25:32<4:09:22, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 460/1801 [1:25:32<4:09:22, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 461/1801 [1:25:42<4:09:09, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 461/1801 [1:25:42<4:09:09, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 462/1801 [1:25:54<4:08:59, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 462/1801 [1:25:54<4:08:59, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 463/1801 [1:26:06<4:08:49, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 463/1801 [1:26:06<4:08:49, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 464/1801 [1:26:17<4:08:39, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 464/1801 [1:26:17<4:08:39, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 465/1801 [1:26:28<4:08:28, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 465/1801 [1:26:28<4:08:28, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 466/1801 [1:26:39<4:08:16, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 466/1801 [1:26:39<4:08:16, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 467/1801 [1:26:51<4:08:06, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 467/1801 [1:26:51<4:08:06, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 468/1801 [1:27:02<4:07:54, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 468/1801 [1:27:02<4:07:54, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 469/1801 [1:27:13<4:07:44, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 469/1801 [1:27:13<4:07:44, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 470/1801 [1:27:24<4:07:32, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 470/1801 [1:27:24<4:07:32, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 471/1801 [1:27:35<4:07:21, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 471/1801 [1:27:35<4:07:21, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 472/1801 [1:27:46<4:07:09, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▌ | 472/1801 [1:27:46<4:07:09, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 473/1801 [1:27:57<4:06:57, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 473/1801 [1:27:57<4:06:57, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 474/1801 [1:28:07<4:06:43, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 474/1801 [1:28:07<4:06:43, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 475/1801 [1:28:18<4:06:30, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 475/1801 [1:28:18<4:06:30, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 476/1801 [1:28:29<4:06:19, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 476/1801 [1:28:29<4:06:19, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 477/1801 [1:28:40<4:06:07, 0.09it/s, v_num=ufl5] Epoch 0: 26%|██▋ | 477/1801 [1:28:40<4:06:07, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 478/1801 [1:28:51<4:05:55, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 478/1801 [1:28:51<4:05:55, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 479/1801 [1:29:02<4:05:43, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 479/1801 [1:29:02<4:05:43, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 480/1801 [1:29:13<4:05:33, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 480/1801 [1:29:13<4:05:33, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 481/1801 [1:29:24<4:05:20, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 481/1801 [1:29:24<4:05:20, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 482/1801 [1:29:34<4:05:07, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 482/1801 [1:29:34<4:05:07, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 483/1801 [1:29:45<4:04:56, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 483/1801 [1:29:45<4:04:56, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 484/1801 [1:29:56<4:04:43, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 484/1801 [1:29:56<4:04:43, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 485/1801 [1:30:06<4:04:30, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 485/1801 [1:30:06<4:04:30, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 486/1801 [1:30:17<4:04:18, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 486/1801 [1:30:17<4:04:18, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 487/1801 [1:30:28<4:04:06, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 487/1801 [1:30:28<4:04:06, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 488/1801 [1:30:40<4:03:58, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 488/1801 [1:30:40<4:03:58, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 489/1801 [1:30:50<4:03:44, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 489/1801 [1:30:50<4:03:44, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 490/1801 [1:31:02<4:03:34, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 490/1801 [1:31:02<4:03:34, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 491/1801 [1:31:13<4:03:22, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 491/1801 [1:31:13<4:03:22, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 492/1801 [1:31:23<4:03:09, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 492/1801 [1:31:23<4:03:09, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 493/1801 [1:31:34<4:02:57, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 493/1801 [1:31:34<4:02:57, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 494/1801 [1:31:45<4:02:45, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 494/1801 [1:31:45<4:02:45, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 495/1801 [1:31:55<4:02:32, 0.09it/s, v_num=ufl5] Epoch 0: 27%|██▋ | 495/1801 [1:31:55<4:02:32, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 496/1801 [1:32:08<4:02:24, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 496/1801 [1:32:08<4:02:24, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 497/1801 [1:32:18<4:02:12, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 497/1801 [1:32:18<4:02:12, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 498/1801 [1:32:29<4:02:00, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 498/1801 [1:32:29<4:02:00, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 499/1801 [1:32:39<4:01:46, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 499/1801 [1:32:39<4:01:46, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 500/1801 [1:32:51<4:01:36, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 500/1801 [1:32:51<4:01:36, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 501/1801 [1:33:01<4:01:23, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 501/1801 [1:33:01<4:01:23, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 502/1801 [1:33:12<4:01:11, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 502/1801 [1:33:12<4:01:11, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 503/1801 [1:33:23<4:01:00, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 503/1801 [1:33:23<4:01:00, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 504/1801 [1:33:35<4:00:50, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 504/1801 [1:33:35<4:00:50, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 505/1801 [1:33:46<4:00:39, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 505/1801 [1:33:46<4:00:39, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 506/1801 [1:33:57<4:00:28, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 506/1801 [1:33:57<4:00:28, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 507/1801 [1:34:09<4:00:17, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 507/1801 [1:34:09<4:00:17, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 508/1801 [1:34:19<4:00:04, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 508/1801 [1:34:19<4:00:04, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 509/1801 [1:34:30<3:59:53, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 509/1801 [1:34:30<3:59:53, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 510/1801 [1:34:41<3:59:42, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 510/1801 [1:34:41<3:59:42, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 511/1801 [1:34:52<3:59:30, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 511/1801 [1:34:52<3:59:30, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 512/1801 [1:35:04<3:59:21, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 512/1801 [1:35:04<3:59:21, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 513/1801 [1:35:15<3:59:09, 0.09it/s, v_num=ufl5] Epoch 0: 28%|██▊ | 513/1801 [1:35:15<3:59:09, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 514/1801 [1:35:25<3:58:56, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 514/1801 [1:35:25<3:58:56, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 515/1801 [1:35:37<3:58:47, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 515/1801 [1:35:37<3:58:47, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 516/1801 [1:35:48<3:58:35, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 516/1801 [1:35:48<3:58:35, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 517/1801 [1:35:59<3:58:25, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▊ | 517/1801 [1:35:59<3:58:25, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 518/1801 [1:36:10<3:58:11, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 518/1801 [1:36:10<3:58:11, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 519/1801 [1:36:20<3:57:58, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 519/1801 [1:36:20<3:57:58, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 520/1801 [1:36:32<3:57:48, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 520/1801 [1:36:32<3:57:48, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 521/1801 [1:36:42<3:57:34, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 521/1801 [1:36:42<3:57:34, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 522/1801 [1:36:53<3:57:25, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 522/1801 [1:36:53<3:57:25, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 523/1801 [1:37:04<3:57:13, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 523/1801 [1:37:04<3:57:13, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 524/1801 [1:37:15<3:57:01, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 524/1801 [1:37:15<3:57:01, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 525/1801 [1:37:25<3:56:48, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 525/1801 [1:37:26<3:56:48, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 526/1801 [1:37:37<3:56:37, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 526/1801 [1:37:37<3:56:37, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 527/1801 [1:37:48<3:56:25, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 527/1801 [1:37:48<3:56:25, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 528/1801 [1:38:00<3:56:16, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 528/1801 [1:38:00<3:56:16, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 529/1801 [1:38:10<3:56:04, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 529/1801 [1:38:10<3:56:04, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 530/1801 [1:38:22<3:55:53, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 530/1801 [1:38:22<3:55:53, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 531/1801 [1:38:32<3:55:41, 0.09it/s, v_num=ufl5] Epoch 0: 29%|██▉ | 531/1801 [1:38:32<3:55:41, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 532/1801 [1:38:43<3:55:30, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 532/1801 [1:38:43<3:55:30, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 533/1801 [1:38:55<3:55:19, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 533/1801 [1:38:55<3:55:19, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 534/1801 [1:39:06<3:55:08, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 534/1801 [1:39:06<3:55:08, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 535/1801 [1:39:17<3:54:56, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 535/1801 [1:39:17<3:54:56, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 536/1801 [1:39:28<3:54:45, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 536/1801 [1:39:28<3:54:45, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 537/1801 [1:39:38<3:54:32, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 537/1801 [1:39:38<3:54:32, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 538/1801 [1:39:49<3:54:20, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 538/1801 [1:39:49<3:54:20, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 539/1801 [1:40:00<3:54:10, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 539/1801 [1:40:00<3:54:10, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 540/1801 [1:40:12<3:53:59, 0.09it/s, v_num=ufl5] Epoch 0: 30%|██▉ | 540/1801 [1:40:12<3:53:59, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 541/1801 [1:40:22<3:53:46, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 541/1801 [1:40:22<3:53:46, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 542/1801 [1:40:33<3:53:34, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 542/1801 [1:40:33<3:53:34, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 543/1801 [1:40:43<3:53:22, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 543/1801 [1:40:43<3:53:22, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 544/1801 [1:40:55<3:53:12, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 544/1801 [1:40:55<3:53:12, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 545/1801 [1:41:05<3:52:59, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 545/1801 [1:41:05<3:52:59, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 546/1801 [1:41:16<3:52:48, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 546/1801 [1:41:16<3:52:48, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 547/1801 [1:41:27<3:52:35, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 547/1801 [1:41:27<3:52:35, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 548/1801 [1:41:38<3:52:25, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 548/1801 [1:41:38<3:52:25, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 549/1801 [1:41:50<3:52:13, 0.09it/s, v_num=ufl5] Epoch 0: 30%|███ | 549/1801 [1:41:50<3:52:13, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 550/1801 [1:42:02<3:52:05, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 550/1801 [1:42:02<3:52:05, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 551/1801 [1:42:12<3:51:52, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 551/1801 [1:42:12<3:51:52, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 552/1801 [1:42:24<3:51:42, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 552/1801 [1:42:24<3:51:42, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 553/1801 [1:42:34<3:51:29, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 553/1801 [1:42:34<3:51:29, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 554/1801 [1:42:45<3:51:17, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 554/1801 [1:42:45<3:51:17, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 555/1801 [1:42:57<3:51:07, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 555/1801 [1:42:57<3:51:07, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 556/1801 [1:43:07<3:50:56, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 556/1801 [1:43:07<3:50:56, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 557/1801 [1:43:19<3:50:45, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 557/1801 [1:43:19<3:50:45, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 558/1801 [1:43:30<3:50:34, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 558/1801 [1:43:30<3:50:34, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 559/1801 [1:43:41<3:50:22, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 559/1801 [1:43:41<3:50:22, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 560/1801 [1:43:52<3:50:12, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 560/1801 [1:43:52<3:50:12, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 561/1801 [1:44:03<3:50:00, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 561/1801 [1:44:03<3:50:00, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 562/1801 [1:44:14<3:49:48, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███ | 562/1801 [1:44:14<3:49:48, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 563/1801 [1:44:25<3:49:38, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 563/1801 [1:44:25<3:49:38, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 564/1801 [1:44:36<3:49:25, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 564/1801 [1:44:36<3:49:25, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 565/1801 [1:44:47<3:49:14, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 565/1801 [1:44:47<3:49:14, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 566/1801 [1:44:58<3:49:02, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 566/1801 [1:44:58<3:49:02, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 567/1801 [1:45:09<3:48:51, 0.09it/s, v_num=ufl5] Epoch 0: 31%|███▏ | 567/1801 [1:45:09<3:48:51, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 568/1801 [1:45:21<3:48:43, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 568/1801 [1:45:21<3:48:43, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 569/1801 [1:45:32<3:48:32, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 569/1801 [1:45:32<3:48:32, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 570/1801 [1:45:43<3:48:20, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 570/1801 [1:45:43<3:48:20, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 571/1801 [1:45:54<3:48:07, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 571/1801 [1:45:54<3:48:07, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 572/1801 [1:46:05<3:47:56, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 572/1801 [1:46:05<3:47:56, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 573/1801 [1:46:16<3:47:44, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 573/1801 [1:46:16<3:47:44, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 574/1801 [1:46:27<3:47:33, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 574/1801 [1:46:27<3:47:33, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 575/1801 [1:46:38<3:47:22, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 575/1801 [1:46:38<3:47:22, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 576/1801 [1:46:49<3:47:11, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 576/1801 [1:46:49<3:47:11, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 577/1801 [1:47:00<3:47:00, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 577/1801 [1:47:00<3:47:00, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 578/1801 [1:47:11<3:46:49, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 578/1801 [1:47:11<3:46:49, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 579/1801 [1:47:23<3:46:38, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 579/1801 [1:47:23<3:46:38, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 580/1801 [1:47:33<3:46:26, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 580/1801 [1:47:33<3:46:26, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 581/1801 [1:47:44<3:46:14, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 581/1801 [1:47:44<3:46:14, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 582/1801 [1:47:55<3:46:02, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 582/1801 [1:47:55<3:46:02, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 583/1801 [1:48:06<3:45:50, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 583/1801 [1:48:06<3:45:50, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 584/1801 [1:48:17<3:45:41, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 584/1801 [1:48:17<3:45:41, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 585/1801 [1:48:28<3:45:29, 0.09it/s, v_num=ufl5] Epoch 0: 32%|███▏ | 585/1801 [1:48:28<3:45:29, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 586/1801 [1:48:38<3:45:16, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 586/1801 [1:48:38<3:45:16, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 587/1801 [1:48:49<3:45:04, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 587/1801 [1:48:49<3:45:04, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 588/1801 [1:49:00<3:44:52, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 588/1801 [1:49:00<3:44:52, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 589/1801 [1:49:11<3:44:41, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 589/1801 [1:49:11<3:44:41, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 590/1801 [1:49:22<3:44:30, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 590/1801 [1:49:22<3:44:30, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 591/1801 [1:49:34<3:44:19, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 591/1801 [1:49:34<3:44:19, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 592/1801 [1:49:46<3:44:10, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 592/1801 [1:49:46<3:44:10, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 593/1801 [1:49:57<3:43:58, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 593/1801 [1:49:57<3:43:58, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 594/1801 [1:50:07<3:43:46, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 594/1801 [1:50:07<3:43:46, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 595/1801 [1:50:18<3:43:35, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 595/1801 [1:50:18<3:43:35, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 596/1801 [1:50:29<3:43:22, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 596/1801 [1:50:29<3:43:22, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 597/1801 [1:50:39<3:43:10, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 597/1801 [1:50:39<3:43:10, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 598/1801 [1:50:51<3:43:00, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 598/1801 [1:50:51<3:43:00, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 599/1801 [1:51:02<3:42:48, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 599/1801 [1:51:02<3:42:48, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 600/1801 [1:51:14<3:42:39, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 600/1801 [1:51:14<3:42:39, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 601/1801 [1:51:24<3:42:27, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 601/1801 [1:51:24<3:42:27, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 602/1801 [1:51:36<3:42:17, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 602/1801 [1:51:36<3:42:17, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 603/1801 [1:51:47<3:42:06, 0.09it/s, v_num=ufl5] Epoch 0: 33%|███▎ | 603/1801 [1:51:47<3:42:06, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 604/1801 [1:51:58<3:41:53, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 604/1801 [1:51:58<3:41:53, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 605/1801 [1:52:09<3:41:42, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 605/1801 [1:52:09<3:41:42, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 606/1801 [1:52:20<3:41:31, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 606/1801 [1:52:20<3:41:31, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 607/1801 [1:52:31<3:41:19, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▎ | 607/1801 [1:52:31<3:41:19, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 608/1801 [1:52:42<3:41:09, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 608/1801 [1:52:42<3:41:09, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 609/1801 [1:52:53<3:40:57, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 609/1801 [1:52:53<3:40:57, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 610/1801 [1:53:04<3:40:46, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 610/1801 [1:53:04<3:40:46, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 611/1801 [1:53:14<3:40:33, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 611/1801 [1:53:14<3:40:33, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 612/1801 [1:53:25<3:40:21, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 612/1801 [1:53:25<3:40:21, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 613/1801 [1:53:35<3:40:08, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 613/1801 [1:53:35<3:40:08, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 614/1801 [1:53:46<3:39:57, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 614/1801 [1:53:46<3:39:57, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 615/1801 [1:53:56<3:39:44, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 615/1801 [1:53:56<3:39:44, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 616/1801 [1:54:09<3:39:35, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 616/1801 [1:54:09<3:39:35, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 617/1801 [1:54:20<3:39:24, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 617/1801 [1:54:20<3:39:24, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 618/1801 [1:54:30<3:39:12, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 618/1801 [1:54:30<3:39:12, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 619/1801 [1:54:41<3:39:00, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 619/1801 [1:54:41<3:39:00, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 620/1801 [1:54:51<3:38:47, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 620/1801 [1:54:51<3:38:47, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 621/1801 [1:55:02<3:38:36, 0.09it/s, v_num=ufl5] Epoch 0: 34%|███▍ | 621/1801 [1:55:02<3:38:36, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 622/1801 [1:55:14<3:38:25, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 622/1801 [1:55:14<3:38:25, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 623/1801 [1:55:24<3:38:13, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 623/1801 [1:55:24<3:38:13, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 624/1801 [1:55:36<3:38:03, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 624/1801 [1:55:36<3:38:03, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 625/1801 [1:55:47<3:37:52, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 625/1801 [1:55:47<3:37:52, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 626/1801 [1:55:58<3:37:40, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 626/1801 [1:55:58<3:37:40, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 627/1801 [1:56:08<3:37:28, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 627/1801 [1:56:08<3:37:28, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 628/1801 [1:56:19<3:37:16, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 628/1801 [1:56:19<3:37:16, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 629/1801 [1:56:30<3:37:05, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 629/1801 [1:56:30<3:37:05, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 630/1801 [1:56:41<3:36:53, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▍ | 630/1801 [1:56:41<3:36:53, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 631/1801 [1:56:51<3:36:41, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 631/1801 [1:56:51<3:36:41, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 632/1801 [1:57:03<3:36:31, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 632/1801 [1:57:03<3:36:31, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 633/1801 [1:57:13<3:36:18, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 633/1801 [1:57:13<3:36:18, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 634/1801 [1:57:24<3:36:06, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 634/1801 [1:57:24<3:36:06, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 635/1801 [1:57:35<3:35:54, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 635/1801 [1:57:35<3:35:54, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 636/1801 [1:57:45<3:35:42, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 636/1801 [1:57:45<3:35:42, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 637/1801 [1:57:56<3:35:30, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 637/1801 [1:57:56<3:35:30, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 638/1801 [1:58:06<3:35:18, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 638/1801 [1:58:06<3:35:18, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 639/1801 [1:58:17<3:35:06, 0.09it/s, v_num=ufl5] Epoch 0: 35%|███▌ | 639/1801 [1:58:17<3:35:06, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 640/1801 [1:58:28<3:34:56, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 640/1801 [1:58:28<3:34:56, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 641/1801 [1:58:39<3:34:43, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 641/1801 [1:58:39<3:34:43, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 642/1801 [1:58:50<3:34:33, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 642/1801 [1:58:50<3:34:33, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 643/1801 [1:59:02<3:34:23, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 643/1801 [1:59:02<3:34:23, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 644/1801 [1:59:13<3:34:11, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 644/1801 [1:59:13<3:34:11, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 645/1801 [1:59:24<3:34:00, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 645/1801 [1:59:24<3:34:00, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 646/1801 [1:59:35<3:33:48, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 646/1801 [1:59:35<3:33:48, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 647/1801 [1:59:45<3:33:36, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 647/1801 [1:59:45<3:33:36, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 648/1801 [1:59:57<3:33:25, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 648/1801 [1:59:57<3:33:25, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 649/1801 [2:00:08<3:33:14, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 649/1801 [2:00:08<3:33:14, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 650/1801 [2:00:19<3:33:03, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 650/1801 [2:00:19<3:33:03, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 651/1801 [2:00:30<3:32:52, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 651/1801 [2:00:30<3:32:52, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 652/1801 [2:00:40<3:32:40, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▌ | 652/1801 [2:00:40<3:32:40, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 653/1801 [2:00:51<3:32:28, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 653/1801 [2:00:51<3:32:28, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 654/1801 [2:01:01<3:32:16, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 654/1801 [2:01:01<3:32:16, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 655/1801 [2:01:12<3:32:04, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 655/1801 [2:01:12<3:32:04, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 656/1801 [2:01:24<3:31:55, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 656/1801 [2:01:24<3:31:55, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 657/1801 [2:01:35<3:31:43, 0.09it/s, v_num=ufl5] Epoch 0: 36%|███▋ | 657/1801 [2:01:35<3:31:43, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 658/1801 [2:01:46<3:31:32, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 658/1801 [2:01:46<3:31:32, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 659/1801 [2:01:57<3:31:20, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 659/1801 [2:01:57<3:31:20, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 660/1801 [2:02:07<3:31:07, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 660/1801 [2:02:07<3:31:07, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 661/1801 [2:02:19<3:30:57, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 661/1801 [2:02:19<3:30:57, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 662/1801 [2:02:29<3:30:45, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 662/1801 [2:02:29<3:30:45, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 663/1801 [2:02:40<3:30:33, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 663/1801 [2:02:40<3:30:33, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 664/1801 [2:02:52<3:30:25, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 664/1801 [2:02:52<3:30:25, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 665/1801 [2:03:03<3:30:13, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 665/1801 [2:03:03<3:30:13, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 666/1801 [2:03:14<3:30:02, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 666/1801 [2:03:14<3:30:02, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 667/1801 [2:03:25<3:29:50, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 667/1801 [2:03:25<3:29:50, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 668/1801 [2:03:36<3:29:39, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 668/1801 [2:03:36<3:29:39, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 669/1801 [2:03:48<3:29:28, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 669/1801 [2:03:48<3:29:28, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 670/1801 [2:03:58<3:29:16, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 670/1801 [2:03:58<3:29:16, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 671/1801 [2:04:09<3:29:05, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 671/1801 [2:04:09<3:29:05, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 672/1801 [2:04:20<3:28:54, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 672/1801 [2:04:20<3:28:54, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 673/1801 [2:04:31<3:28:43, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 673/1801 [2:04:31<3:28:43, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 674/1801 [2:04:42<3:28:30, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 674/1801 [2:04:42<3:28:30, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 675/1801 [2:04:53<3:28:20, 0.09it/s, v_num=ufl5] Epoch 0: 37%|███▋ | 675/1801 [2:04:53<3:28:20, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 676/1801 [2:05:04<3:28:08, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 676/1801 [2:05:04<3:28:08, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 677/1801 [2:05:14<3:27:56, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 677/1801 [2:05:14<3:27:56, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 678/1801 [2:05:25<3:27:44, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 678/1801 [2:05:25<3:27:44, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 679/1801 [2:05:36<3:27:33, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 679/1801 [2:05:36<3:27:33, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 680/1801 [2:05:47<3:27:22, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 680/1801 [2:05:47<3:27:22, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 681/1801 [2:05:58<3:27:11, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 681/1801 [2:05:58<3:27:11, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 682/1801 [2:06:10<3:27:00, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 682/1801 [2:06:10<3:27:00, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 683/1801 [2:06:20<3:26:49, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 683/1801 [2:06:20<3:26:49, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 684/1801 [2:06:32<3:26:38, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 684/1801 [2:06:32<3:26:38, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 685/1801 [2:06:42<3:26:25, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 685/1801 [2:06:42<3:26:25, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 686/1801 [2:06:53<3:26:14, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 686/1801 [2:06:53<3:26:14, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 687/1801 [2:07:04<3:26:03, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 687/1801 [2:07:04<3:26:03, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 688/1801 [2:07:16<3:25:53, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 688/1801 [2:07:16<3:25:53, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 689/1801 [2:07:27<3:25:42, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 689/1801 [2:07:27<3:25:42, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 690/1801 [2:07:39<3:25:32, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 690/1801 [2:07:39<3:25:32, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 691/1801 [2:07:49<3:25:20, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 691/1801 [2:07:49<3:25:20, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 692/1801 [2:08:01<3:25:10, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 692/1801 [2:08:01<3:25:10, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 693/1801 [2:08:12<3:24:58, 0.09it/s, v_num=ufl5] Epoch 0: 38%|███▊ | 693/1801 [2:08:12<3:24:58, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 694/1801 [2:08:23<3:24:47, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 694/1801 [2:08:23<3:24:47, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 695/1801 [2:08:34<3:24:37, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 695/1801 [2:08:34<3:24:37, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 696/1801 [2:08:46<3:24:27, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 696/1801 [2:08:46<3:24:27, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 697/1801 [2:08:57<3:24:15, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▊ | 697/1801 [2:08:57<3:24:15, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 698/1801 [2:09:07<3:24:03, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 698/1801 [2:09:07<3:24:03, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 699/1801 [2:09:18<3:23:51, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 699/1801 [2:09:18<3:23:51, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 700/1801 [2:09:29<3:23:40, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 700/1801 [2:09:29<3:23:40, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 701/1801 [2:09:40<3:23:28, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 701/1801 [2:09:40<3:23:28, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 702/1801 [2:09:51<3:23:17, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 702/1801 [2:09:51<3:23:17, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 703/1801 [2:10:02<3:23:06, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 703/1801 [2:10:02<3:23:06, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 704/1801 [2:10:14<3:22:56, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 704/1801 [2:10:14<3:22:56, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 705/1801 [2:10:25<3:22:45, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 705/1801 [2:10:25<3:22:45, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 706/1801 [2:10:36<3:22:34, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 706/1801 [2:10:36<3:22:34, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 707/1801 [2:10:48<3:22:24, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 707/1801 [2:10:48<3:22:24, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 708/1801 [2:10:59<3:22:13, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 708/1801 [2:10:59<3:22:13, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 709/1801 [2:11:10<3:22:01, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 709/1801 [2:11:10<3:22:01, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 710/1801 [2:11:21<3:21:50, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 710/1801 [2:11:21<3:21:50, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 711/1801 [2:11:32<3:21:40, 0.09it/s, v_num=ufl5] Epoch 0: 39%|███▉ | 711/1801 [2:11:32<3:21:40, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 712/1801 [2:11:44<3:21:29, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 712/1801 [2:11:44<3:21:29, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 713/1801 [2:11:55<3:21:18, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 713/1801 [2:11:55<3:21:18, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 714/1801 [2:12:05<3:21:06, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 714/1801 [2:12:05<3:21:06, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 715/1801 [2:12:16<3:20:55, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 715/1801 [2:12:16<3:20:55, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 716/1801 [2:12:27<3:20:43, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 716/1801 [2:12:27<3:20:43, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 717/1801 [2:12:38<3:20:31, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 717/1801 [2:12:38<3:20:31, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 718/1801 [2:12:49<3:20:21, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 718/1801 [2:12:49<3:20:21, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 719/1801 [2:13:00<3:20:09, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 719/1801 [2:13:00<3:20:09, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 720/1801 [2:13:12<3:19:59, 0.09it/s, v_num=ufl5] Epoch 0: 40%|███▉ | 720/1801 [2:13:12<3:19:59, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 721/1801 [2:13:23<3:19:49, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 721/1801 [2:13:23<3:19:49, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 722/1801 [2:13:34<3:19:37, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 722/1801 [2:13:34<3:19:37, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 723/1801 [2:13:45<3:19:26, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 723/1801 [2:13:45<3:19:26, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 724/1801 [2:13:55<3:19:13, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 724/1801 [2:13:55<3:19:13, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 725/1801 [2:14:06<3:19:02, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 725/1801 [2:14:06<3:19:02, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 726/1801 [2:14:17<3:18:50, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 726/1801 [2:14:17<3:18:50, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 727/1801 [2:14:27<3:18:38, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 727/1801 [2:14:27<3:18:38, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 728/1801 [2:14:39<3:18:28, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 728/1801 [2:14:39<3:18:28, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 729/1801 [2:14:50<3:18:17, 0.09it/s, v_num=ufl5] Epoch 0: 40%|████ | 729/1801 [2:14:50<3:18:17, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 730/1801 [2:15:01<3:18:06, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 730/1801 [2:15:01<3:18:06, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 731/1801 [2:15:11<3:17:53, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 731/1801 [2:15:11<3:17:53, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 732/1801 [2:15:22<3:17:42, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 732/1801 [2:15:22<3:17:42, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 733/1801 [2:15:33<3:17:30, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 733/1801 [2:15:33<3:17:30, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 734/1801 [2:15:43<3:17:18, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 734/1801 [2:15:43<3:17:18, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 735/1801 [2:15:54<3:17:07, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 735/1801 [2:15:54<3:17:07, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 736/1801 [2:16:06<3:16:56, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 736/1801 [2:16:06<3:16:56, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 737/1801 [2:16:17<3:16:45, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 737/1801 [2:16:17<3:16:45, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 738/1801 [2:16:27<3:16:33, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 738/1801 [2:16:27<3:16:33, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 739/1801 [2:16:38<3:16:21, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 739/1801 [2:16:38<3:16:21, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 740/1801 [2:16:49<3:16:10, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 740/1801 [2:16:49<3:16:10, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 741/1801 [2:17:00<3:15:58, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 741/1801 [2:17:00<3:15:58, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 742/1801 [2:17:10<3:15:46, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████ | 742/1801 [2:17:10<3:15:46, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 743/1801 [2:17:20<3:15:34, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 743/1801 [2:17:20<3:15:34, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 744/1801 [2:17:32<3:15:24, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 744/1801 [2:17:32<3:15:24, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 745/1801 [2:17:43<3:15:13, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 745/1801 [2:17:43<3:15:13, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 746/1801 [2:17:54<3:15:01, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 746/1801 [2:17:54<3:15:01, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 747/1801 [2:18:04<3:14:49, 0.09it/s, v_num=ufl5] Epoch 0: 41%|████▏ | 747/1801 [2:18:04<3:14:49, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 748/1801 [2:18:14<3:14:37, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 748/1801 [2:18:14<3:14:37, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 749/1801 [2:18:25<3:14:25, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 749/1801 [2:18:25<3:14:25, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 750/1801 [2:18:36<3:14:14, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 750/1801 [2:18:36<3:14:14, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 751/1801 [2:18:46<3:14:02, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 751/1801 [2:18:46<3:14:02, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 752/1801 [2:18:59<3:13:52, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 752/1801 [2:18:59<3:13:52, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 753/1801 [2:19:09<3:13:40, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 753/1801 [2:19:09<3:13:40, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 754/1801 [2:19:19<3:13:28, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 754/1801 [2:19:19<3:13:28, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 755/1801 [2:19:29<3:13:15, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 755/1801 [2:19:29<3:13:15, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 756/1801 [2:19:40<3:13:03, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 756/1801 [2:19:40<3:13:03, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 757/1801 [2:19:51<3:12:52, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 757/1801 [2:19:51<3:12:52, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 758/1801 [2:20:02<3:12:41, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 758/1801 [2:20:02<3:12:41, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 759/1801 [2:20:13<3:12:30, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 759/1801 [2:20:13<3:12:30, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 760/1801 [2:20:24<3:12:19, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 760/1801 [2:20:24<3:12:19, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 761/1801 [2:20:35<3:12:08, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 761/1801 [2:20:35<3:12:08, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 762/1801 [2:20:45<3:11:56, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 762/1801 [2:20:45<3:11:56, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 763/1801 [2:20:56<3:11:43, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 763/1801 [2:20:56<3:11:43, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 764/1801 [2:21:06<3:11:31, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 764/1801 [2:21:06<3:11:31, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 765/1801 [2:21:16<3:11:19, 0.09it/s, v_num=ufl5] Epoch 0: 42%|████▏ | 765/1801 [2:21:16<3:11:19, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 766/1801 [2:21:27<3:11:08, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 766/1801 [2:21:27<3:11:08, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 767/1801 [2:21:38<3:10:57, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 767/1801 [2:21:38<3:10:57, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 768/1801 [2:21:51<3:10:47, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 768/1801 [2:21:51<3:10:47, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 769/1801 [2:22:01<3:10:36, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 769/1801 [2:22:01<3:10:36, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 770/1801 [2:22:13<3:10:25, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 770/1801 [2:22:13<3:10:25, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 771/1801 [2:22:23<3:10:13, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 771/1801 [2:22:23<3:10:13, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 772/1801 [2:22:33<3:10:01, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 772/1801 [2:22:33<3:10:01, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 773/1801 [2:22:44<3:09:50, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 773/1801 [2:22:44<3:09:50, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 774/1801 [2:22:56<3:09:39, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 774/1801 [2:22:56<3:09:39, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 775/1801 [2:23:06<3:09:28, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 775/1801 [2:23:06<3:09:28, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 776/1801 [2:23:18<3:09:17, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 776/1801 [2:23:18<3:09:17, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 777/1801 [2:23:29<3:09:06, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 777/1801 [2:23:29<3:09:06, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 778/1801 [2:23:40<3:08:54, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 778/1801 [2:23:40<3:08:54, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 779/1801 [2:23:50<3:08:42, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 779/1801 [2:23:50<3:08:42, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 780/1801 [2:24:01<3:08:31, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 780/1801 [2:24:01<3:08:31, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 781/1801 [2:24:11<3:08:19, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 781/1801 [2:24:11<3:08:19, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 782/1801 [2:24:23<3:08:08, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 782/1801 [2:24:23<3:08:08, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 783/1801 [2:24:33<3:07:57, 0.09it/s, v_num=ufl5] Epoch 0: 43%|████▎ | 783/1801 [2:24:33<3:07:57, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 784/1801 [2:24:45<3:07:47, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 784/1801 [2:24:45<3:07:47, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 785/1801 [2:24:55<3:07:34, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 785/1801 [2:24:55<3:07:34, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 786/1801 [2:25:06<3:07:22, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 786/1801 [2:25:06<3:07:22, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 787/1801 [2:25:17<3:07:11, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▎ | 787/1801 [2:25:17<3:07:11, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 788/1801 [2:25:28<3:07:00, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 788/1801 [2:25:28<3:07:00, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 789/1801 [2:25:39<3:06:49, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 789/1801 [2:25:39<3:06:49, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 790/1801 [2:25:49<3:06:37, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 790/1801 [2:25:49<3:06:37, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 791/1801 [2:25:59<3:06:25, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 791/1801 [2:25:59<3:06:25, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 792/1801 [2:26:11<3:06:14, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 792/1801 [2:26:11<3:06:14, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 793/1801 [2:26:22<3:06:03, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 793/1801 [2:26:22<3:06:03, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 794/1801 [2:26:33<3:05:52, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 794/1801 [2:26:33<3:05:52, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 795/1801 [2:26:44<3:05:41, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 795/1801 [2:26:44<3:05:41, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 796/1801 [2:26:55<3:05:30, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 796/1801 [2:26:55<3:05:30, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 797/1801 [2:27:06<3:05:19, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 797/1801 [2:27:06<3:05:19, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 798/1801 [2:27:17<3:05:07, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 798/1801 [2:27:17<3:05:07, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 799/1801 [2:27:28<3:04:55, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 799/1801 [2:27:28<3:04:55, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 800/1801 [2:27:40<3:04:46, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 800/1801 [2:27:40<3:04:46, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 801/1801 [2:27:50<3:04:33, 0.09it/s, v_num=ufl5] Epoch 0: 44%|████▍ | 801/1801 [2:27:50<3:04:33, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 802/1801 [2:28:01<3:04:22, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 802/1801 [2:28:01<3:04:22, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 803/1801 [2:28:11<3:04:10, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 803/1801 [2:28:11<3:04:10, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 804/1801 [2:28:22<3:03:59, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 804/1801 [2:28:22<3:03:59, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 805/1801 [2:28:32<3:03:47, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 805/1801 [2:28:32<3:03:47, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 806/1801 [2:28:43<3:03:36, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 806/1801 [2:28:43<3:03:36, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 807/1801 [2:28:54<3:03:24, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 807/1801 [2:28:54<3:03:24, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 808/1801 [2:29:06<3:03:15, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 808/1801 [2:29:06<3:03:15, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 809/1801 [2:29:17<3:03:03, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 809/1801 [2:29:17<3:03:03, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 810/1801 [2:29:28<3:02:52, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▍ | 810/1801 [2:29:28<3:02:52, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 811/1801 [2:29:39<3:02:41, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 811/1801 [2:29:39<3:02:41, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 812/1801 [2:29:50<3:02:29, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 812/1801 [2:29:50<3:02:29, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 813/1801 [2:30:00<3:02:18, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 813/1801 [2:30:00<3:02:18, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 814/1801 [2:30:11<3:02:07, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 814/1801 [2:30:11<3:02:07, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 815/1801 [2:30:22<3:01:55, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 815/1801 [2:30:22<3:01:55, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 816/1801 [2:30:34<3:01:45, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 816/1801 [2:30:34<3:01:45, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 817/1801 [2:30:46<3:01:35, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 817/1801 [2:30:46<3:01:35, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 818/1801 [2:30:57<3:01:24, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 818/1801 [2:30:57<3:01:24, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 819/1801 [2:31:08<3:01:13, 0.09it/s, v_num=ufl5] Epoch 0: 45%|████▌ | 819/1801 [2:31:08<3:01:13, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 820/1801 [2:31:20<3:01:02, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 820/1801 [2:31:20<3:01:02, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 821/1801 [2:31:31<3:00:52, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 821/1801 [2:31:31<3:00:52, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 822/1801 [2:31:43<3:00:42, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 822/1801 [2:31:43<3:00:42, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 823/1801 [2:31:55<3:00:32, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 823/1801 [2:31:55<3:00:32, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 824/1801 [2:32:08<3:00:23, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 824/1801 [2:32:08<3:00:23, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 825/1801 [2:32:20<3:00:12, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 825/1801 [2:32:20<3:00:12, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 826/1801 [2:32:31<3:00:02, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 826/1801 [2:32:31<3:00:02, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 827/1801 [2:32:42<2:59:51, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 827/1801 [2:32:42<2:59:51, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 828/1801 [2:32:53<2:59:40, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 828/1801 [2:32:53<2:59:40, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 829/1801 [2:33:04<2:59:29, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 829/1801 [2:33:05<2:59:29, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 830/1801 [2:33:17<2:59:19, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 830/1801 [2:33:17<2:59:19, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 831/1801 [2:33:28<2:59:08, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 831/1801 [2:33:28<2:59:08, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 832/1801 [2:33:40<2:58:58, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▌ | 832/1801 [2:33:40<2:58:58, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 833/1801 [2:33:51<2:58:47, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 833/1801 [2:33:51<2:58:47, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 834/1801 [2:34:03<2:58:37, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 834/1801 [2:34:03<2:58:37, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 835/1801 [2:34:13<2:58:25, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 835/1801 [2:34:13<2:58:25, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 836/1801 [2:34:25<2:58:15, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 836/1801 [2:34:25<2:58:15, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 837/1801 [2:34:36<2:58:03, 0.09it/s, v_num=ufl5] Epoch 0: 46%|████▋ | 837/1801 [2:34:36<2:58:03, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 838/1801 [2:34:47<2:57:52, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 838/1801 [2:34:47<2:57:52, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 839/1801 [2:34:58<2:57:42, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 839/1801 [2:34:58<2:57:42, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 840/1801 [2:35:11<2:57:32, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 840/1801 [2:35:11<2:57:32, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 841/1801 [2:35:22<2:57:21, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 841/1801 [2:35:22<2:57:21, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 842/1801 [2:35:33<2:57:10, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 842/1801 [2:35:33<2:57:10, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 843/1801 [2:35:44<2:56:58, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 843/1801 [2:35:44<2:56:58, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 844/1801 [2:35:54<2:56:47, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 844/1801 [2:35:54<2:56:47, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 845/1801 [2:36:06<2:56:36, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 845/1801 [2:36:06<2:56:36, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 846/1801 [2:36:17<2:56:25, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 846/1801 [2:36:17<2:56:25, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 847/1801 [2:36:28<2:56:14, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 847/1801 [2:36:28<2:56:14, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 848/1801 [2:36:40<2:56:04, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 848/1801 [2:36:40<2:56:04, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 849/1801 [2:36:52<2:55:54, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 849/1801 [2:36:52<2:55:54, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 850/1801 [2:37:04<2:55:43, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 850/1801 [2:37:04<2:55:44, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 851/1801 [2:37:15<2:55:33, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 851/1801 [2:37:15<2:55:33, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 852/1801 [2:37:26<2:55:22, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 852/1801 [2:37:26<2:55:22, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 853/1801 [2:37:37<2:55:11, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 853/1801 [2:37:37<2:55:11, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 854/1801 [2:37:48<2:55:00, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 854/1801 [2:37:48<2:55:00, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 855/1801 [2:38:00<2:54:49, 0.09it/s, v_num=ufl5] Epoch 0: 47%|████▋ | 855/1801 [2:38:00<2:54:49, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 856/1801 [2:38:12<2:54:39, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 856/1801 [2:38:12<2:54:39, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 857/1801 [2:38:23<2:54:28, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 857/1801 [2:38:23<2:54:28, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 858/1801 [2:38:35<2:54:17, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 858/1801 [2:38:35<2:54:17, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 859/1801 [2:38:46<2:54:06, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 859/1801 [2:38:46<2:54:06, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 860/1801 [2:38:57<2:53:55, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 860/1801 [2:38:57<2:53:55, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 861/1801 [2:39:08<2:53:44, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 861/1801 [2:39:08<2:53:44, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 862/1801 [2:39:18<2:53:32, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 862/1801 [2:39:18<2:53:32, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 863/1801 [2:39:29<2:53:20, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 863/1801 [2:39:29<2:53:20, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 864/1801 [2:39:41<2:53:10, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 864/1801 [2:39:41<2:53:10, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 865/1801 [2:39:51<2:52:59, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 865/1801 [2:39:51<2:52:59, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 866/1801 [2:40:02<2:52:47, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 866/1801 [2:40:02<2:52:47, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 867/1801 [2:40:13<2:52:36, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 867/1801 [2:40:13<2:52:36, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 868/1801 [2:40:24<2:52:25, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 868/1801 [2:40:24<2:52:25, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 869/1801 [2:40:35<2:52:14, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 869/1801 [2:40:35<2:52:14, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 870/1801 [2:40:46<2:52:02, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 870/1801 [2:40:46<2:52:02, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 871/1801 [2:40:57<2:51:52, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 871/1801 [2:40:57<2:51:52, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 872/1801 [2:41:10<2:51:42, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 872/1801 [2:41:10<2:51:42, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 873/1801 [2:41:21<2:51:30, 0.09it/s, v_num=ufl5] Epoch 0: 48%|████▊ | 873/1801 [2:41:21<2:51:30, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 874/1801 [2:41:32<2:51:20, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 874/1801 [2:41:32<2:51:20, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 875/1801 [2:41:43<2:51:09, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 875/1801 [2:41:43<2:51:09, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 876/1801 [2:41:54<2:50:58, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 876/1801 [2:41:54<2:50:58, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 877/1801 [2:42:05<2:50:46, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▊ | 877/1801 [2:42:05<2:50:46, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 878/1801 [2:42:15<2:50:34, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 878/1801 [2:42:15<2:50:34, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 879/1801 [2:42:26<2:50:23, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 879/1801 [2:42:26<2:50:23, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 880/1801 [2:42:38<2:50:12, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 880/1801 [2:42:38<2:50:12, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 881/1801 [2:42:48<2:50:01, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 881/1801 [2:42:48<2:50:01, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 882/1801 [2:43:00<2:49:50, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 882/1801 [2:43:00<2:49:50, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 883/1801 [2:43:11<2:49:39, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 883/1801 [2:43:11<2:49:39, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 884/1801 [2:43:21<2:49:27, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 884/1801 [2:43:21<2:49:27, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 885/1801 [2:43:32<2:49:16, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 885/1801 [2:43:32<2:49:16, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 886/1801 [2:43:43<2:49:05, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 886/1801 [2:43:43<2:49:05, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 887/1801 [2:43:54<2:48:53, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 887/1801 [2:43:54<2:48:53, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 888/1801 [2:44:06<2:48:43, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 888/1801 [2:44:06<2:48:43, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 889/1801 [2:44:17<2:48:32, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 889/1801 [2:44:17<2:48:32, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 890/1801 [2:44:27<2:48:20, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 890/1801 [2:44:27<2:48:20, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 891/1801 [2:44:39<2:48:09, 0.09it/s, v_num=ufl5] Epoch 0: 49%|████▉ | 891/1801 [2:44:39<2:48:09, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 892/1801 [2:44:50<2:47:58, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 892/1801 [2:44:50<2:47:58, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 893/1801 [2:45:00<2:47:47, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 893/1801 [2:45:00<2:47:47, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 894/1801 [2:45:11<2:47:35, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 894/1801 [2:45:11<2:47:35, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 895/1801 [2:45:22<2:47:24, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 895/1801 [2:45:22<2:47:24, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 896/1801 [2:45:33<2:47:13, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 896/1801 [2:45:33<2:47:13, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 897/1801 [2:45:44<2:47:02, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 897/1801 [2:45:44<2:47:02, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 898/1801 [2:45:55<2:46:50, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 898/1801 [2:45:55<2:46:50, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 899/1801 [2:46:05<2:46:38, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 899/1801 [2:46:05<2:46:38, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 900/1801 [2:46:16<2:46:27, 0.09it/s, v_num=ufl5] Epoch 0: 50%|████▉ | 900/1801 [2:46:16<2:46:27, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 901/1801 [2:46:26<2:46:15, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 901/1801 [2:46:26<2:46:15, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 902/1801 [2:46:37<2:46:03, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 902/1801 [2:46:37<2:46:03, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 903/1801 [2:46:48<2:45:52, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 903/1801 [2:46:48<2:45:52, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 904/1801 [2:47:00<2:45:43, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 904/1801 [2:47:00<2:45:43, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 905/1801 [2:47:11<2:45:32, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 905/1801 [2:47:11<2:45:32, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 906/1801 [2:47:22<2:45:20, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 906/1801 [2:47:22<2:45:20, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 907/1801 [2:47:33<2:45:09, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 907/1801 [2:47:33<2:45:09, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 908/1801 [2:47:44<2:44:58, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 908/1801 [2:47:44<2:44:58, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 909/1801 [2:47:55<2:44:46, 0.09it/s, v_num=ufl5] Epoch 0: 50%|█████ | 909/1801 [2:47:55<2:44:46, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 910/1801 [2:48:06<2:44:35, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 910/1801 [2:48:06<2:44:35, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 911/1801 [2:48:16<2:44:23, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 911/1801 [2:48:16<2:44:23, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 912/1801 [2:48:29<2:44:14, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 912/1801 [2:48:29<2:44:14, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 913/1801 [2:48:40<2:44:03, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 913/1801 [2:48:40<2:44:03, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 914/1801 [2:48:51<2:43:52, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 914/1801 [2:48:51<2:43:52, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 915/1801 [2:49:02<2:43:41, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 915/1801 [2:49:02<2:43:41, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 916/1801 [2:49:13<2:43:29, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 916/1801 [2:49:13<2:43:29, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 917/1801 [2:49:24<2:43:19, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 917/1801 [2:49:24<2:43:19, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 918/1801 [2:49:35<2:43:07, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 918/1801 [2:49:35<2:43:07, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 919/1801 [2:49:46<2:42:56, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 919/1801 [2:49:46<2:42:56, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 920/1801 [2:49:58<2:42:45, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 920/1801 [2:49:58<2:42:45, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 921/1801 [2:50:08<2:42:34, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 921/1801 [2:50:08<2:42:34, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 922/1801 [2:50:19<2:42:23, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 922/1801 [2:50:19<2:42:23, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 923/1801 [2:50:30<2:42:12, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████ | 923/1801 [2:50:30<2:42:12, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 924/1801 [2:50:41<2:42:00, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 924/1801 [2:50:41<2:42:00, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 925/1801 [2:50:52<2:41:49, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 925/1801 [2:50:52<2:41:49, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 926/1801 [2:51:02<2:41:37, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 926/1801 [2:51:02<2:41:37, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 927/1801 [2:51:13<2:41:26, 0.09it/s, v_num=ufl5] Epoch 0: 51%|█████▏ | 927/1801 [2:51:13<2:41:26, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 928/1801 [2:51:25<2:41:15, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 928/1801 [2:51:25<2:41:15, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 929/1801 [2:51:35<2:41:04, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 929/1801 [2:51:35<2:41:04, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 930/1801 [2:51:47<2:40:53, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 930/1801 [2:51:47<2:40:53, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 931/1801 [2:51:58<2:40:42, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 931/1801 [2:51:58<2:40:42, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 932/1801 [2:52:09<2:40:31, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 932/1801 [2:52:09<2:40:31, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 933/1801 [2:52:21<2:40:20, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 933/1801 [2:52:21<2:40:20, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 934/1801 [2:52:32<2:40:09, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 934/1801 [2:52:32<2:40:09, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 935/1801 [2:52:43<2:39:58, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 935/1801 [2:52:43<2:39:58, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 936/1801 [2:52:55<2:39:48, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 936/1801 [2:52:55<2:39:48, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 937/1801 [2:53:06<2:39:37, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 937/1801 [2:53:06<2:39:37, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 938/1801 [2:53:17<2:39:26, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 938/1801 [2:53:17<2:39:26, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 939/1801 [2:53:27<2:39:14, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 939/1801 [2:53:27<2:39:14, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 940/1801 [2:53:39<2:39:03, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 940/1801 [2:53:39<2:39:03, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 941/1801 [2:53:49<2:38:52, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 941/1801 [2:53:49<2:38:52, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 942/1801 [2:54:01<2:38:41, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 942/1801 [2:54:01<2:38:41, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 943/1801 [2:54:11<2:38:29, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 943/1801 [2:54:11<2:38:29, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 944/1801 [2:54:22<2:38:18, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 944/1801 [2:54:22<2:38:18, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 945/1801 [2:54:33<2:38:07, 0.09it/s, v_num=ufl5] Epoch 0: 52%|█████▏ | 945/1801 [2:54:33<2:38:07, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 946/1801 [2:54:44<2:37:55, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 946/1801 [2:54:44<2:37:55, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 947/1801 [2:54:55<2:37:44, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 947/1801 [2:54:55<2:37:44, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 948/1801 [2:55:06<2:37:33, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 948/1801 [2:55:06<2:37:33, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 949/1801 [2:55:17<2:37:22, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 949/1801 [2:55:17<2:37:22, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 950/1801 [2:55:28<2:37:10, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 950/1801 [2:55:28<2:37:10, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 951/1801 [2:55:39<2:36:59, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 951/1801 [2:55:39<2:36:59, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 952/1801 [2:55:50<2:36:49, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 952/1801 [2:55:50<2:36:49, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 953/1801 [2:56:01<2:36:37, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 953/1801 [2:56:01<2:36:37, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 954/1801 [2:56:12<2:36:26, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 954/1801 [2:56:12<2:36:26, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 955/1801 [2:56:22<2:36:14, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 955/1801 [2:56:22<2:36:14, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 956/1801 [2:56:33<2:36:03, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 956/1801 [2:56:33<2:36:03, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 957/1801 [2:56:44<2:35:52, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 957/1801 [2:56:44<2:35:52, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 958/1801 [2:56:54<2:35:40, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 958/1801 [2:56:54<2:35:40, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 959/1801 [2:57:05<2:35:28, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 959/1801 [2:57:05<2:35:28, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 960/1801 [2:57:17<2:35:18, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 960/1801 [2:57:17<2:35:18, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 961/1801 [2:57:27<2:35:06, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 961/1801 [2:57:27<2:35:06, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 962/1801 [2:57:37<2:34:55, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 962/1801 [2:57:37<2:34:55, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 963/1801 [2:57:48<2:34:43, 0.09it/s, v_num=ufl5] Epoch 0: 53%|█████▎ | 963/1801 [2:57:48<2:34:43, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 964/1801 [2:57:58<2:34:31, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 964/1801 [2:57:58<2:34:31, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 965/1801 [2:58:09<2:34:20, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 965/1801 [2:58:09<2:34:20, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 966/1801 [2:58:19<2:34:08, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 966/1801 [2:58:19<2:34:08, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 967/1801 [2:58:30<2:33:57, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 967/1801 [2:58:30<2:33:57, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 968/1801 [2:58:42<2:33:46, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▎ | 968/1801 [2:58:42<2:33:46, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 969/1801 [2:58:52<2:33:35, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 969/1801 [2:58:52<2:33:35, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 970/1801 [2:59:03<2:33:24, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 970/1801 [2:59:03<2:33:24, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 971/1801 [2:59:14<2:33:12, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 971/1801 [2:59:14<2:33:12, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 972/1801 [2:59:25<2:33:01, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 972/1801 [2:59:25<2:33:01, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 973/1801 [2:59:37<2:32:51, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 973/1801 [2:59:37<2:32:51, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 974/1801 [2:59:48<2:32:39, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 974/1801 [2:59:48<2:32:39, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 975/1801 [2:59:58<2:32:28, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 975/1801 [2:59:58<2:32:28, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 976/1801 [3:00:10<2:32:17, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 976/1801 [3:00:10<2:32:17, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 977/1801 [3:00:21<2:32:06, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 977/1801 [3:00:21<2:32:06, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 978/1801 [3:00:32<2:31:55, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 978/1801 [3:00:32<2:31:55, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 979/1801 [3:00:43<2:31:44, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 979/1801 [3:00:43<2:31:44, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 980/1801 [3:00:54<2:31:33, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 980/1801 [3:00:54<2:31:33, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 981/1801 [3:01:04<2:31:21, 0.09it/s, v_num=ufl5] Epoch 0: 54%|█████▍ | 981/1801 [3:01:04<2:31:21, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 982/1801 [3:01:15<2:31:10, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 982/1801 [3:01:15<2:31:10, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 983/1801 [3:01:26<2:30:59, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 983/1801 [3:01:26<2:30:59, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 984/1801 [3:01:38<2:30:48, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 984/1801 [3:01:38<2:30:48, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 985/1801 [3:01:49<2:30:37, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 985/1801 [3:01:49<2:30:37, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 986/1801 [3:02:00<2:30:26, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 986/1801 [3:02:00<2:30:26, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 987/1801 [3:02:10<2:30:14, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 987/1801 [3:02:10<2:30:14, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 988/1801 [3:02:21<2:30:03, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 988/1801 [3:02:21<2:30:03, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 989/1801 [3:02:32<2:29:52, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 989/1801 [3:02:32<2:29:52, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 990/1801 [3:02:43<2:29:41, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▍ | 990/1801 [3:02:43<2:29:41, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 991/1801 [3:02:54<2:29:30, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 991/1801 [3:02:54<2:29:30, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 992/1801 [3:03:06<2:29:19, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 992/1801 [3:03:06<2:29:19, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 993/1801 [3:03:17<2:29:08, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 993/1801 [3:03:17<2:29:08, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 994/1801 [3:03:27<2:28:56, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 994/1801 [3:03:27<2:28:56, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 995/1801 [3:03:38<2:28:45, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 995/1801 [3:03:38<2:28:45, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 996/1801 [3:03:49<2:28:34, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 996/1801 [3:03:49<2:28:34, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 997/1801 [3:03:59<2:28:22, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 997/1801 [3:03:59<2:28:22, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 998/1801 [3:04:09<2:28:10, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 998/1801 [3:04:09<2:28:10, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 999/1801 [3:04:20<2:27:59, 0.09it/s, v_num=ufl5] Epoch 0: 55%|█████▌ | 999/1801 [3:04:20<2:27:59, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1000/1801 [3:04:31<2:27:48, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1000/1801 [3:04:31<2:27:48, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1001/1801 [3:04:42<2:27:37, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1001/1801 [3:04:42<2:27:37, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1002/1801 [3:04:53<2:27:26, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1002/1801 [3:04:53<2:27:26, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1003/1801 [3:05:04<2:27:14, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1003/1801 [3:05:04<2:27:14, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1004/1801 [3:05:15<2:27:03, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1004/1801 [3:05:15<2:27:03, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1005/1801 [3:05:26<2:26:52, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1005/1801 [3:05:26<2:26:52, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1006/1801 [3:05:37<2:26:41, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1006/1801 [3:05:37<2:26:41, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1007/1801 [3:05:47<2:26:29, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1007/1801 [3:05:47<2:26:29, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1008/1801 [3:05:59<2:26:19, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1008/1801 [3:05:59<2:26:19, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1009/1801 [3:06:09<2:26:07, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1009/1801 [3:06:09<2:26:07, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1010/1801 [3:06:20<2:25:56, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1010/1801 [3:06:20<2:25:56, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1011/1801 [3:06:32<2:25:45, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1011/1801 [3:06:32<2:25:45, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1012/1801 [3:06:42<2:25:33, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1012/1801 [3:06:42<2:25:33, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1013/1801 [3:06:53<2:25:22, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▌ | 1013/1801 [3:06:53<2:25:22, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1014/1801 [3:07:04<2:25:11, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1014/1801 [3:07:04<2:25:11, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1015/1801 [3:07:14<2:24:59, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1015/1801 [3:07:14<2:24:59, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1016/1801 [3:07:26<2:24:49, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1016/1801 [3:07:26<2:24:49, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1017/1801 [3:07:37<2:24:38, 0.09it/s, v_num=ufl5] Epoch 0: 56%|█████▋ | 1017/1801 [3:07:37<2:24:38, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1018/1801 [3:07:47<2:24:26, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1018/1801 [3:07:47<2:24:26, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1019/1801 [3:07:58<2:24:15, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1019/1801 [3:07:58<2:24:15, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1020/1801 [3:08:09<2:24:04, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1020/1801 [3:08:09<2:24:04, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1021/1801 [3:08:20<2:23:53, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1021/1801 [3:08:20<2:23:53, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1022/1801 [3:08:30<2:23:41, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1022/1801 [3:08:30<2:23:41, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1023/1801 [3:08:41<2:23:30, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1023/1801 [3:08:41<2:23:30, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1024/1801 [3:08:53<2:23:19, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1024/1801 [3:08:53<2:23:19, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1025/1801 [3:09:04<2:23:08, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1025/1801 [3:09:04<2:23:08, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1026/1801 [3:09:14<2:22:57, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1026/1801 [3:09:14<2:22:57, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1027/1801 [3:09:25<2:22:45, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1027/1801 [3:09:25<2:22:45, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1028/1801 [3:09:36<2:22:34, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1028/1801 [3:09:36<2:22:34, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1029/1801 [3:09:47<2:22:23, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1029/1801 [3:09:47<2:22:23, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1030/1801 [3:09:57<2:22:11, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1030/1801 [3:09:57<2:22:11, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1031/1801 [3:10:09<2:22:00, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1031/1801 [3:10:09<2:22:00, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1032/1801 [3:10:21<2:21:50, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1032/1801 [3:10:21<2:21:50, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1033/1801 [3:10:32<2:21:39, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1033/1801 [3:10:32<2:21:39, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1034/1801 [3:10:43<2:21:28, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1034/1801 [3:10:43<2:21:28, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1035/1801 [3:10:54<2:21:17, 0.09it/s, v_num=ufl5] Epoch 0: 57%|█████▋ | 1035/1801 [3:10:54<2:21:17, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1036/1801 [3:11:05<2:21:06, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1036/1801 [3:11:05<2:21:06, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1037/1801 [3:11:15<2:20:54, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1037/1801 [3:11:15<2:20:54, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1038/1801 [3:11:26<2:20:43, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1038/1801 [3:11:26<2:20:43, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1039/1801 [3:11:36<2:20:31, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1039/1801 [3:11:36<2:20:31, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1040/1801 [3:11:48<2:20:20, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1040/1801 [3:11:48<2:20:20, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1041/1801 [3:11:59<2:20:10, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1041/1801 [3:11:59<2:20:10, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1042/1801 [3:12:09<2:19:58, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1042/1801 [3:12:09<2:19:58, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1043/1801 [3:12:20<2:19:47, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1043/1801 [3:12:20<2:19:47, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1044/1801 [3:12:31<2:19:35, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1044/1801 [3:12:31<2:19:35, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1045/1801 [3:12:42<2:19:24, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1045/1801 [3:12:42<2:19:24, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1046/1801 [3:12:53<2:19:13, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1046/1801 [3:12:53<2:19:13, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1047/1801 [3:13:04<2:19:02, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1047/1801 [3:13:04<2:19:02, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1048/1801 [3:13:16<2:18:52, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1048/1801 [3:13:16<2:18:52, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1049/1801 [3:13:27<2:18:41, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1049/1801 [3:13:27<2:18:41, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1050/1801 [3:13:38<2:18:29, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1050/1801 [3:13:38<2:18:29, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1051/1801 [3:13:48<2:18:18, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1051/1801 [3:13:48<2:18:18, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1052/1801 [3:13:58<2:18:06, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1052/1801 [3:13:58<2:18:06, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1053/1801 [3:14:09<2:17:55, 0.09it/s, v_num=ufl5] Epoch 0: 58%|█████▊ | 1053/1801 [3:14:09<2:17:55, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1054/1801 [3:14:19<2:17:43, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1054/1801 [3:14:19<2:17:43, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1055/1801 [3:14:31<2:17:32, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1055/1801 [3:14:31<2:17:32, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1056/1801 [3:14:42<2:17:21, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1056/1801 [3:14:42<2:17:21, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1057/1801 [3:14:54<2:17:11, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1057/1801 [3:14:54<2:17:11, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1058/1801 [3:15:04<2:16:59, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▊ | 1058/1801 [3:15:04<2:16:59, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1059/1801 [3:15:16<2:16:49, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1059/1801 [3:15:16<2:16:49, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1060/1801 [3:15:27<2:16:37, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1060/1801 [3:15:27<2:16:37, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1061/1801 [3:15:37<2:16:26, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1061/1801 [3:15:37<2:16:26, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1062/1801 [3:15:49<2:16:16, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1062/1801 [3:15:49<2:16:16, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1063/1801 [3:16:00<2:16:05, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1063/1801 [3:16:00<2:16:05, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1064/1801 [3:16:13<2:15:55, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1064/1801 [3:16:13<2:15:55, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1065/1801 [3:16:24<2:15:44, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1065/1801 [3:16:24<2:15:44, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1066/1801 [3:16:35<2:15:32, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1066/1801 [3:16:35<2:15:32, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1067/1801 [3:16:45<2:15:21, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1067/1801 [3:16:45<2:15:21, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1068/1801 [3:16:56<2:15:10, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1068/1801 [3:16:56<2:15:10, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1069/1801 [3:17:07<2:14:59, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1069/1801 [3:17:07<2:14:59, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1070/1801 [3:17:18<2:14:48, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1070/1801 [3:17:18<2:14:48, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1071/1801 [3:17:29<2:14:36, 0.09it/s, v_num=ufl5] Epoch 0: 59%|█████▉ | 1071/1801 [3:17:29<2:14:36, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1072/1801 [3:17:41<2:14:26, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1072/1801 [3:17:41<2:14:26, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1073/1801 [3:17:52<2:14:14, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1073/1801 [3:17:52<2:14:14, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1074/1801 [3:18:03<2:14:03, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1074/1801 [3:18:03<2:14:03, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1075/1801 [3:18:14<2:13:52, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1075/1801 [3:18:14<2:13:52, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1076/1801 [3:18:25<2:13:41, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1076/1801 [3:18:25<2:13:41, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1077/1801 [3:18:36<2:13:30, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1077/1801 [3:18:36<2:13:30, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1078/1801 [3:18:47<2:13:19, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1078/1801 [3:18:47<2:13:19, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1079/1801 [3:18:57<2:13:08, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1079/1801 [3:18:57<2:13:08, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1080/1801 [3:19:09<2:12:57, 0.09it/s, v_num=ufl5] Epoch 0: 60%|█████▉ | 1080/1801 [3:19:09<2:12:57, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1081/1801 [3:19:20<2:12:46, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1081/1801 [3:19:20<2:12:46, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1082/1801 [3:19:31<2:12:34, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1082/1801 [3:19:31<2:12:34, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1083/1801 [3:19:41<2:12:23, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1083/1801 [3:19:41<2:12:23, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1084/1801 [3:19:53<2:12:12, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1084/1801 [3:19:53<2:12:12, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1085/1801 [3:20:04<2:12:01, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1085/1801 [3:20:04<2:12:01, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1086/1801 [3:20:15<2:11:50, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1086/1801 [3:20:15<2:11:50, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1087/1801 [3:20:26<2:11:39, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1087/1801 [3:20:26<2:11:39, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1088/1801 [3:20:38<2:11:29, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1088/1801 [3:20:38<2:11:29, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1089/1801 [3:20:49<2:11:18, 0.09it/s, v_num=ufl5] Epoch 0: 60%|██████ | 1089/1801 [3:20:49<2:11:18, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1090/1801 [3:21:00<2:11:06, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1090/1801 [3:21:00<2:11:06, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1091/1801 [3:21:10<2:10:55, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1091/1801 [3:21:10<2:10:55, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1092/1801 [3:21:21<2:10:43, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1092/1801 [3:21:21<2:10:43, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1093/1801 [3:21:32<2:10:33, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1093/1801 [3:21:32<2:10:33, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1094/1801 [3:21:43<2:10:21, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1094/1801 [3:21:43<2:10:21, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1095/1801 [3:21:53<2:10:10, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1095/1801 [3:21:53<2:10:10, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1096/1801 [3:22:05<2:09:59, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1096/1801 [3:22:05<2:09:59, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1097/1801 [3:22:16<2:09:48, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1097/1801 [3:22:16<2:09:48, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1098/1801 [3:22:28<2:09:37, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1098/1801 [3:22:28<2:09:37, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1099/1801 [3:22:38<2:09:26, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1099/1801 [3:22:38<2:09:26, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1100/1801 [3:22:49<2:09:15, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1100/1801 [3:22:49<2:09:15, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1101/1801 [3:23:00<2:09:04, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1101/1801 [3:23:00<2:09:04, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1102/1801 [3:23:11<2:08:52, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1102/1801 [3:23:11<2:08:52, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1103/1801 [3:23:21<2:08:41, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████ | 1103/1801 [3:23:21<2:08:41, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1104/1801 [3:23:33<2:08:30, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1104/1801 [3:23:33<2:08:30, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1105/1801 [3:23:44<2:08:20, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1105/1801 [3:23:44<2:08:20, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1106/1801 [3:23:56<2:08:09, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1106/1801 [3:23:56<2:08:09, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1107/1801 [3:24:07<2:07:58, 0.09it/s, v_num=ufl5] Epoch 0: 61%|██████▏ | 1107/1801 [3:24:07<2:07:58, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1108/1801 [3:24:18<2:07:47, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1108/1801 [3:24:18<2:07:47, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1109/1801 [3:24:29<2:07:35, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1109/1801 [3:24:29<2:07:35, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1110/1801 [3:24:40<2:07:24, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1110/1801 [3:24:40<2:07:24, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1111/1801 [3:24:51<2:07:13, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1111/1801 [3:24:51<2:07:13, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1112/1801 [3:25:03<2:07:03, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1112/1801 [3:25:03<2:07:03, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1113/1801 [3:25:14<2:06:51, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1113/1801 [3:25:14<2:06:51, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1114/1801 [3:25:25<2:06:41, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1114/1801 [3:25:25<2:06:41, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1115/1801 [3:25:35<2:06:29, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1115/1801 [3:25:35<2:06:29, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1116/1801 [3:25:47<2:06:18, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1116/1801 [3:25:47<2:06:18, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1117/1801 [3:25:58<2:06:07, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1117/1801 [3:25:58<2:06:07, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1118/1801 [3:26:09<2:05:56, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1118/1801 [3:26:09<2:05:56, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1119/1801 [3:26:20<2:05:45, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1119/1801 [3:26:20<2:05:45, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1120/1801 [3:26:32<2:05:35, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1120/1801 [3:26:32<2:05:35, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1121/1801 [3:26:43<2:05:23, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1121/1801 [3:26:43<2:05:23, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1122/1801 [3:26:53<2:05:12, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1122/1801 [3:26:53<2:05:12, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1123/1801 [3:27:04<2:05:01, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1123/1801 [3:27:04<2:05:01, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1124/1801 [3:27:15<2:04:50, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1124/1801 [3:27:15<2:04:50, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1125/1801 [3:27:27<2:04:39, 0.09it/s, v_num=ufl5] Epoch 0: 62%|██████▏ | 1125/1801 [3:27:27<2:04:39, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1126/1801 [3:27:38<2:04:28, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1126/1801 [3:27:38<2:04:28, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1127/1801 [3:27:48<2:04:16, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1127/1801 [3:27:48<2:04:16, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1128/1801 [3:27:59<2:04:05, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1128/1801 [3:27:59<2:04:05, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1129/1801 [3:28:10<2:03:54, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1129/1801 [3:28:10<2:03:54, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1130/1801 [3:28:22<2:03:43, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1130/1801 [3:28:22<2:03:43, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1131/1801 [3:28:32<2:03:32, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1131/1801 [3:28:32<2:03:32, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1132/1801 [3:28:44<2:03:21, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1132/1801 [3:28:44<2:03:21, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1133/1801 [3:28:55<2:03:10, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1133/1801 [3:28:55<2:03:10, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1134/1801 [3:29:06<2:02:59, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1134/1801 [3:29:06<2:02:59, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1135/1801 [3:29:18<2:02:49, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1135/1801 [3:29:18<2:02:49, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1136/1801 [3:29:30<2:02:38, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1136/1801 [3:29:30<2:02:38, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1137/1801 [3:29:42<2:02:28, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1137/1801 [3:29:42<2:02:28, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1138/1801 [3:29:53<2:02:17, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1138/1801 [3:29:53<2:02:17, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1139/1801 [3:30:04<2:02:06, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1139/1801 [3:30:04<2:02:06, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1140/1801 [3:30:15<2:01:54, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1140/1801 [3:30:15<2:01:54, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1141/1801 [3:30:26<2:01:43, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1141/1801 [3:30:26<2:01:43, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1142/1801 [3:30:38<2:01:33, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1142/1801 [3:30:38<2:01:33, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1143/1801 [3:30:49<2:01:22, 0.09it/s, v_num=ufl5] Epoch 0: 63%|██████▎ | 1143/1801 [3:30:49<2:01:22, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1144/1801 [3:31:01<2:01:11, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1144/1801 [3:31:01<2:01:11, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1145/1801 [3:31:13<2:01:00, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1145/1801 [3:31:13<2:01:00, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1146/1801 [3:31:24<2:00:49, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1146/1801 [3:31:24<2:00:49, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1147/1801 [3:31:35<2:00:38, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1147/1801 [3:31:35<2:00:38, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1148/1801 [3:31:45<2:00:27, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▎ | 1148/1801 [3:31:45<2:00:27, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1149/1801 [3:31:56<2:00:16, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1149/1801 [3:31:56<2:00:16, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1150/1801 [3:32:08<2:00:05, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1150/1801 [3:32:08<2:00:05, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1151/1801 [3:32:18<1:59:53, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1151/1801 [3:32:18<1:59:53, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1152/1801 [3:32:29<1:59:42, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1152/1801 [3:32:29<1:59:42, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1153/1801 [3:32:41<1:59:31, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1153/1801 [3:32:41<1:59:31, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1154/1801 [3:32:51<1:59:20, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1154/1801 [3:32:51<1:59:20, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1155/1801 [3:33:01<1:59:08, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1155/1801 [3:33:01<1:59:08, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1156/1801 [3:33:13<1:58:58, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1156/1801 [3:33:13<1:58:58, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1157/1801 [3:33:24<1:58:47, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1157/1801 [3:33:24<1:58:47, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1158/1801 [3:33:35<1:58:36, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1158/1801 [3:33:35<1:58:36, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1159/1801 [3:33:47<1:58:25, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1159/1801 [3:33:47<1:58:25, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1160/1801 [3:33:59<1:58:15, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1160/1801 [3:33:59<1:58:15, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1161/1801 [3:34:11<1:58:04, 0.09it/s, v_num=ufl5] Epoch 0: 64%|██████▍ | 1161/1801 [3:34:11<1:58:04, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1162/1801 [3:34:23<1:57:53, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1162/1801 [3:34:23<1:57:53, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1163/1801 [3:34:34<1:57:42, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1163/1801 [3:34:34<1:57:42, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1164/1801 [3:34:45<1:57:31, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1164/1801 [3:34:45<1:57:31, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1165/1801 [3:34:56<1:57:20, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1165/1801 [3:34:56<1:57:20, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1166/1801 [3:35:07<1:57:09, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1166/1801 [3:35:07<1:57:09, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1167/1801 [3:35:18<1:56:58, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1167/1801 [3:35:18<1:56:58, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1168/1801 [3:35:30<1:56:47, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1168/1801 [3:35:30<1:56:47, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1169/1801 [3:35:40<1:56:36, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1169/1801 [3:35:40<1:56:36, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1170/1801 [3:35:51<1:56:24, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▍ | 1170/1801 [3:35:51<1:56:24, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1171/1801 [3:36:02<1:56:13, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1171/1801 [3:36:02<1:56:13, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1172/1801 [3:36:13<1:56:02, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1172/1801 [3:36:13<1:56:02, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1173/1801 [3:36:24<1:55:51, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1173/1801 [3:36:24<1:55:51, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1174/1801 [3:36:35<1:55:40, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1174/1801 [3:36:35<1:55:40, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1175/1801 [3:36:45<1:55:29, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1175/1801 [3:36:45<1:55:29, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1176/1801 [3:36:57<1:55:18, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1176/1801 [3:36:57<1:55:18, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1177/1801 [3:37:09<1:55:07, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1177/1801 [3:37:09<1:55:07, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1178/1801 [3:37:20<1:54:56, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1178/1801 [3:37:20<1:54:56, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1179/1801 [3:37:30<1:54:45, 0.09it/s, v_num=ufl5] Epoch 0: 65%|██████▌ | 1179/1801 [3:37:30<1:54:45, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1180/1801 [3:37:42<1:54:34, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1180/1801 [3:37:42<1:54:34, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1181/1801 [3:37:53<1:54:23, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1181/1801 [3:37:53<1:54:23, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1182/1801 [3:38:03<1:54:11, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1182/1801 [3:38:03<1:54:11, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1183/1801 [3:38:14<1:54:00, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1183/1801 [3:38:14<1:54:00, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1184/1801 [3:38:26<1:53:49, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1184/1801 [3:38:26<1:53:49, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1185/1801 [3:38:37<1:53:38, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1185/1801 [3:38:37<1:53:38, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1186/1801 [3:38:48<1:53:27, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1186/1801 [3:38:48<1:53:27, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1187/1801 [3:39:00<1:53:17, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1187/1801 [3:39:00<1:53:17, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1188/1801 [3:39:10<1:53:05, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1188/1801 [3:39:10<1:53:05, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1189/1801 [3:39:21<1:52:54, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1189/1801 [3:39:21<1:52:54, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1190/1801 [3:39:33<1:52:43, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1190/1801 [3:39:33<1:52:43, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1191/1801 [3:39:44<1:52:32, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1191/1801 [3:39:44<1:52:32, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1192/1801 [3:39:56<1:52:22, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1192/1801 [3:39:56<1:52:22, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1193/1801 [3:40:07<1:52:11, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▌ | 1193/1801 [3:40:07<1:52:11, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1194/1801 [3:40:19<1:52:00, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1194/1801 [3:40:19<1:52:00, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1195/1801 [3:40:31<1:51:50, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1195/1801 [3:40:31<1:51:50, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1196/1801 [3:40:42<1:51:38, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1196/1801 [3:40:42<1:51:38, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1197/1801 [3:40:53<1:51:27, 0.09it/s, v_num=ufl5] Epoch 0: 66%|██████▋ | 1197/1801 [3:40:53<1:51:27, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1198/1801 [3:41:05<1:51:16, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1198/1801 [3:41:05<1:51:16, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1199/1801 [3:41:16<1:51:06, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1199/1801 [3:41:16<1:51:06, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1200/1801 [3:41:28<1:50:55, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1200/1801 [3:41:28<1:50:55, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1201/1801 [3:41:39<1:50:44, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1201/1801 [3:41:39<1:50:44, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1202/1801 [3:41:50<1:50:32, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1202/1801 [3:41:50<1:50:32, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1203/1801 [3:42:00<1:50:21, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1203/1801 [3:42:00<1:50:21, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1204/1801 [3:42:11<1:50:10, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1204/1801 [3:42:11<1:50:10, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1205/1801 [3:42:23<1:49:59, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1205/1801 [3:42:23<1:49:59, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1206/1801 [3:42:35<1:49:49, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1206/1801 [3:42:35<1:49:49, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1207/1801 [3:42:46<1:49:37, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1207/1801 [3:42:46<1:49:37, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1208/1801 [3:42:57<1:49:27, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1208/1801 [3:42:57<1:49:27, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1209/1801 [3:43:08<1:49:16, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1209/1801 [3:43:08<1:49:16, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1210/1801 [3:43:19<1:49:04, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1210/1801 [3:43:19<1:49:04, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1211/1801 [3:43:30<1:48:53, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1211/1801 [3:43:30<1:48:53, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1212/1801 [3:43:41<1:48:42, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1212/1801 [3:43:41<1:48:42, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1213/1801 [3:43:52<1:48:31, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1213/1801 [3:43:52<1:48:31, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1214/1801 [3:44:04<1:48:20, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1214/1801 [3:44:04<1:48:20, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1215/1801 [3:44:15<1:48:09, 0.09it/s, v_num=ufl5] Epoch 0: 67%|██████▋ | 1215/1801 [3:44:15<1:48:09, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1216/1801 [3:44:27<1:47:59, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1216/1801 [3:44:27<1:47:59, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1217/1801 [3:44:38<1:47:48, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1217/1801 [3:44:38<1:47:48, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1218/1801 [3:44:50<1:47:37, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1218/1801 [3:44:50<1:47:37, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1219/1801 [3:45:01<1:47:26, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1219/1801 [3:45:01<1:47:26, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1220/1801 [3:45:12<1:47:15, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1220/1801 [3:45:12<1:47:15, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1221/1801 [3:45:23<1:47:03, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1221/1801 [3:45:23<1:47:03, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1222/1801 [3:45:34<1:46:52, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1222/1801 [3:45:34<1:46:52, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1223/1801 [3:45:45<1:46:41, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1223/1801 [3:45:45<1:46:41, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1224/1801 [3:45:56<1:46:30, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1224/1801 [3:45:56<1:46:30, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1225/1801 [3:46:08<1:46:19, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1225/1801 [3:46:08<1:46:19, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1226/1801 [3:46:18<1:46:08, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1226/1801 [3:46:18<1:46:08, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1227/1801 [3:46:29<1:45:57, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1227/1801 [3:46:29<1:45:57, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1228/1801 [3:46:40<1:45:46, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1228/1801 [3:46:40<1:45:46, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1229/1801 [3:46:51<1:45:34, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1229/1801 [3:46:51<1:45:34, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1230/1801 [3:47:02<1:45:23, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1230/1801 [3:47:02<1:45:23, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1231/1801 [3:47:12<1:45:12, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1231/1801 [3:47:12<1:45:12, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1232/1801 [3:47:24<1:45:01, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1232/1801 [3:47:24<1:45:01, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1233/1801 [3:47:36<1:44:50, 0.09it/s, v_num=ufl5] Epoch 0: 68%|██████▊ | 1233/1801 [3:47:36<1:44:50, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1234/1801 [3:47:47<1:44:39, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1234/1801 [3:47:47<1:44:39, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1235/1801 [3:47:57<1:44:28, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1235/1801 [3:47:57<1:44:28, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1236/1801 [3:48:08<1:44:17, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1236/1801 [3:48:08<1:44:17, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1237/1801 [3:48:20<1:44:06, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1237/1801 [3:48:20<1:44:06, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1238/1801 [3:48:30<1:43:55, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▊ | 1238/1801 [3:48:30<1:43:55, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1239/1801 [3:48:42<1:43:44, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1239/1801 [3:48:42<1:43:44, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1240/1801 [3:48:54<1:43:33, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1240/1801 [3:48:54<1:43:33, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1241/1801 [3:49:05<1:43:22, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1241/1801 [3:49:05<1:43:22, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1242/1801 [3:49:16<1:43:11, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1242/1801 [3:49:16<1:43:11, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1243/1801 [3:49:27<1:43:00, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1243/1801 [3:49:27<1:43:00, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1244/1801 [3:49:38<1:42:49, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1244/1801 [3:49:38<1:42:49, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1245/1801 [3:49:50<1:42:38, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1245/1801 [3:49:50<1:42:38, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1246/1801 [3:50:01<1:42:27, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1246/1801 [3:50:01<1:42:27, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1247/1801 [3:50:12<1:42:16, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1247/1801 [3:50:12<1:42:16, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1248/1801 [3:50:24<1:42:05, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1248/1801 [3:50:24<1:42:05, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1249/1801 [3:50:35<1:41:54, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1249/1801 [3:50:35<1:41:54, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1250/1801 [3:50:45<1:41:43, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1250/1801 [3:50:45<1:41:43, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1251/1801 [3:50:56<1:41:31, 0.09it/s, v_num=ufl5] Epoch 0: 69%|██████▉ | 1251/1801 [3:50:56<1:41:31, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1252/1801 [3:51:07<1:41:20, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1252/1801 [3:51:07<1:41:20, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1253/1801 [3:51:18<1:41:09, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1253/1801 [3:51:18<1:41:09, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1254/1801 [3:51:29<1:40:58, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1254/1801 [3:51:29<1:40:58, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1255/1801 [3:51:41<1:40:47, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1255/1801 [3:51:41<1:40:47, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1256/1801 [3:51:53<1:40:37, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1256/1801 [3:51:53<1:40:37, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1257/1801 [3:52:03<1:40:25, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1257/1801 [3:52:03<1:40:25, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1258/1801 [3:52:15<1:40:15, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1258/1801 [3:52:15<1:40:15, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1259/1801 [3:52:26<1:40:03, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1259/1801 [3:52:26<1:40:03, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1260/1801 [3:52:37<1:39:52, 0.09it/s, v_num=ufl5] Epoch 0: 70%|██████▉ | 1260/1801 [3:52:37<1:39:52, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1261/1801 [3:52:48<1:39:41, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1261/1801 [3:52:48<1:39:41, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1262/1801 [3:52:59<1:39:30, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1262/1801 [3:52:59<1:39:30, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1263/1801 [3:53:10<1:39:19, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1263/1801 [3:53:10<1:39:19, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1264/1801 [3:53:23<1:39:09, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1264/1801 [3:53:23<1:39:09, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1265/1801 [3:53:34<1:38:58, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1265/1801 [3:53:34<1:38:58, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1266/1801 [3:53:45<1:38:46, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1266/1801 [3:53:45<1:38:46, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1267/1801 [3:53:55<1:38:35, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1267/1801 [3:53:55<1:38:35, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1268/1801 [3:54:07<1:38:24, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1268/1801 [3:54:07<1:38:24, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1269/1801 [3:54:18<1:38:13, 0.09it/s, v_num=ufl5] Epoch 0: 70%|███████ | 1269/1801 [3:54:18<1:38:13, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1270/1801 [3:54:28<1:38:02, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1270/1801 [3:54:28<1:38:02, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1271/1801 [3:54:40<1:37:51, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1271/1801 [3:54:40<1:37:51, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1272/1801 [3:54:52<1:37:40, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1272/1801 [3:54:52<1:37:40, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1273/1801 [3:55:02<1:37:29, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1273/1801 [3:55:02<1:37:29, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1274/1801 [3:55:12<1:37:17, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1274/1801 [3:55:12<1:37:17, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1275/1801 [3:55:24<1:37:06, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1275/1801 [3:55:24<1:37:06, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1276/1801 [3:55:34<1:36:55, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1276/1801 [3:55:34<1:36:55, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1277/1801 [3:55:45<1:36:44, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1277/1801 [3:55:45<1:36:44, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1278/1801 [3:55:57<1:36:33, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1278/1801 [3:55:57<1:36:33, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1279/1801 [3:56:08<1:36:22, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1279/1801 [3:56:08<1:36:22, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1280/1801 [3:56:20<1:36:11, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1280/1801 [3:56:20<1:36:11, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1281/1801 [3:56:30<1:36:00, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1281/1801 [3:56:30<1:36:00, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1282/1801 [3:56:42<1:35:49, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1282/1801 [3:56:42<1:35:49, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1283/1801 [3:56:52<1:35:38, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████ | 1283/1801 [3:56:52<1:35:38, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1284/1801 [3:57:04<1:35:27, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1284/1801 [3:57:04<1:35:27, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1285/1801 [3:57:15<1:35:16, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1285/1801 [3:57:15<1:35:16, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1286/1801 [3:57:26<1:35:05, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1286/1801 [3:57:26<1:35:05, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1287/1801 [3:57:37<1:34:53, 0.09it/s, v_num=ufl5] Epoch 0: 71%|███████▏ | 1287/1801 [3:57:37<1:34:54, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1288/1801 [3:57:49<1:34:43, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1288/1801 [3:57:49<1:34:43, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1289/1801 [3:57:59<1:34:32, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1289/1801 [3:57:59<1:34:32, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1290/1801 [3:58:10<1:34:20, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1290/1801 [3:58:10<1:34:20, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1291/1801 [3:58:21<1:34:09, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1291/1801 [3:58:21<1:34:09, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1292/1801 [3:58:32<1:33:58, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1292/1801 [3:58:32<1:33:58, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1293/1801 [3:58:42<1:33:47, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1293/1801 [3:58:42<1:33:47, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1294/1801 [3:58:54<1:33:36, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1294/1801 [3:58:54<1:33:36, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1295/1801 [3:59:05<1:33:25, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1295/1801 [3:59:05<1:33:25, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1296/1801 [3:59:17<1:33:14, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1296/1801 [3:59:17<1:33:14, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1297/1801 [3:59:28<1:33:03, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1297/1801 [3:59:28<1:33:03, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1298/1801 [3:59:40<1:32:52, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1298/1801 [3:59:40<1:32:52, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1299/1801 [3:59:51<1:32:41, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1299/1801 [3:59:51<1:32:41, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1300/1801 [4:00:02<1:32:30, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1300/1801 [4:00:02<1:32:30, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1301/1801 [4:00:13<1:32:19, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1301/1801 [4:00:13<1:32:19, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1302/1801 [4:00:24<1:32:08, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1302/1801 [4:00:24<1:32:08, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1303/1801 [4:00:35<1:31:57, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1303/1801 [4:00:35<1:31:57, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1304/1801 [4:00:48<1:31:46, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1304/1801 [4:00:48<1:31:46, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1305/1801 [4:01:00<1:31:35, 0.09it/s, v_num=ufl5] Epoch 0: 72%|███████▏ | 1305/1801 [4:01:00<1:31:35, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1306/1801 [4:01:11<1:31:24, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1306/1801 [4:01:11<1:31:24, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1307/1801 [4:01:21<1:31:13, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1307/1801 [4:01:21<1:31:13, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1308/1801 [4:01:32<1:31:02, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1308/1801 [4:01:32<1:31:02, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1309/1801 [4:01:43<1:30:51, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1309/1801 [4:01:43<1:30:51, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1310/1801 [4:01:54<1:30:40, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1310/1801 [4:01:54<1:30:40, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1311/1801 [4:02:05<1:30:29, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1311/1801 [4:02:05<1:30:29, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1312/1801 [4:02:17<1:30:18, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1312/1801 [4:02:17<1:30:18, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1313/1801 [4:02:28<1:30:07, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1313/1801 [4:02:28<1:30:07, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1314/1801 [4:02:38<1:29:55, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1314/1801 [4:02:38<1:29:55, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1315/1801 [4:02:50<1:29:44, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1315/1801 [4:02:50<1:29:44, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1316/1801 [4:03:01<1:29:33, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1316/1801 [4:03:01<1:29:33, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1317/1801 [4:03:13<1:29:23, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1317/1801 [4:03:13<1:29:23, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1318/1801 [4:03:23<1:29:11, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1318/1801 [4:03:23<1:29:11, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1319/1801 [4:03:35<1:29:00, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1319/1801 [4:03:35<1:29:00, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1320/1801 [4:03:47<1:28:50, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1320/1801 [4:03:47<1:28:50, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1321/1801 [4:03:58<1:28:38, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1321/1801 [4:03:58<1:28:38, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1322/1801 [4:04:09<1:28:27, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1322/1801 [4:04:09<1:28:27, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1323/1801 [4:04:20<1:28:16, 0.09it/s, v_num=ufl5] Epoch 0: 73%|███████▎ | 1323/1801 [4:04:20<1:28:16, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1324/1801 [4:04:32<1:28:05, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1324/1801 [4:04:32<1:28:05, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1325/1801 [4:04:42<1:27:54, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1325/1801 [4:04:42<1:27:54, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1326/1801 [4:04:53<1:27:43, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1326/1801 [4:04:53<1:27:43, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1327/1801 [4:05:04<1:27:32, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1327/1801 [4:05:04<1:27:32, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1328/1801 [4:05:17<1:27:21, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▎ | 1328/1801 [4:05:17<1:27:21, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1329/1801 [4:05:27<1:27:10, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1329/1801 [4:05:27<1:27:10, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1330/1801 [4:05:38<1:26:59, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1330/1801 [4:05:38<1:26:59, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1331/1801 [4:05:50<1:26:48, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1331/1801 [4:05:50<1:26:48, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1332/1801 [4:06:00<1:26:37, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1332/1801 [4:06:00<1:26:37, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1333/1801 [4:06:12<1:26:26, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1333/1801 [4:06:12<1:26:26, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1334/1801 [4:06:23<1:26:15, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1334/1801 [4:06:23<1:26:15, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1335/1801 [4:06:34<1:26:04, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1335/1801 [4:06:34<1:26:04, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1336/1801 [4:06:47<1:25:53, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1336/1801 [4:06:47<1:25:53, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1337/1801 [4:06:57<1:25:42, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1337/1801 [4:06:57<1:25:42, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1338/1801 [4:07:09<1:25:31, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1338/1801 [4:07:09<1:25:31, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1339/1801 [4:07:20<1:25:20, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1339/1801 [4:07:20<1:25:20, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1340/1801 [4:07:32<1:25:09, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1340/1801 [4:07:32<1:25:09, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1341/1801 [4:07:44<1:24:58, 0.09it/s, v_num=ufl5] Epoch 0: 74%|███████▍ | 1341/1801 [4:07:44<1:24:58, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1342/1801 [4:07:55<1:24:47, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1342/1801 [4:07:55<1:24:47, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1343/1801 [4:08:07<1:24:36, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1343/1801 [4:08:07<1:24:36, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1344/1801 [4:08:18<1:24:26, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1344/1801 [4:08:18<1:24:26, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1345/1801 [4:08:29<1:24:14, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1345/1801 [4:08:29<1:24:14, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1346/1801 [4:08:39<1:24:03, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1346/1801 [4:08:39<1:24:03, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1347/1801 [4:08:50<1:23:52, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1347/1801 [4:08:50<1:23:52, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1348/1801 [4:09:01<1:23:41, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1348/1801 [4:09:01<1:23:41, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1349/1801 [4:09:12<1:23:30, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1349/1801 [4:09:12<1:23:30, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1350/1801 [4:09:23<1:23:18, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▍ | 1350/1801 [4:09:23<1:23:18, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1351/1801 [4:09:34<1:23:07, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1351/1801 [4:09:34<1:23:07, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1352/1801 [4:09:45<1:22:56, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1352/1801 [4:09:45<1:22:56, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1353/1801 [4:09:57<1:22:45, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1353/1801 [4:09:57<1:22:45, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1354/1801 [4:10:08<1:22:34, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1354/1801 [4:10:08<1:22:34, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1355/1801 [4:10:18<1:22:23, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1355/1801 [4:10:18<1:22:23, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1356/1801 [4:10:29<1:22:12, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1356/1801 [4:10:29<1:22:12, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1357/1801 [4:10:41<1:22:01, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1357/1801 [4:10:41<1:22:01, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1358/1801 [4:10:52<1:21:50, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1358/1801 [4:10:52<1:21:50, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1359/1801 [4:11:03<1:21:39, 0.09it/s, v_num=ufl5] Epoch 0: 75%|███████▌ | 1359/1801 [4:11:03<1:21:39, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1360/1801 [4:11:15<1:21:28, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1360/1801 [4:11:15<1:21:28, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1361/1801 [4:11:26<1:21:17, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1361/1801 [4:11:26<1:21:17, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1362/1801 [4:11:38<1:21:06, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1362/1801 [4:11:38<1:21:06, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1363/1801 [4:11:49<1:20:55, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1363/1801 [4:11:49<1:20:55, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1364/1801 [4:12:00<1:20:44, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1364/1801 [4:12:00<1:20:44, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1365/1801 [4:12:11<1:20:33, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1365/1801 [4:12:11<1:20:33, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1366/1801 [4:12:23<1:20:22, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1366/1801 [4:12:23<1:20:22, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1367/1801 [4:12:35<1:20:11, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1367/1801 [4:12:35<1:20:11, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1368/1801 [4:12:46<1:20:00, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1368/1801 [4:12:46<1:20:00, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1369/1801 [4:12:57<1:19:49, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1369/1801 [4:12:57<1:19:49, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1370/1801 [4:13:08<1:19:38, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1370/1801 [4:13:08<1:19:38, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1371/1801 [4:13:19<1:19:27, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1371/1801 [4:13:19<1:19:27, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1372/1801 [4:13:30<1:19:16, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1372/1801 [4:13:30<1:19:16, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1373/1801 [4:13:41<1:19:04, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▌ | 1373/1801 [4:13:41<1:19:04, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1374/1801 [4:13:53<1:18:54, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1374/1801 [4:13:53<1:18:54, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1375/1801 [4:14:04<1:18:43, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1375/1801 [4:14:04<1:18:43, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1376/1801 [4:14:16<1:18:32, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1376/1801 [4:14:16<1:18:32, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1377/1801 [4:14:28<1:18:21, 0.09it/s, v_num=ufl5] Epoch 0: 76%|███████▋ | 1377/1801 [4:14:28<1:18:21, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1378/1801 [4:14:39<1:18:10, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1378/1801 [4:14:39<1:18:10, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1379/1801 [4:14:50<1:17:59, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1379/1801 [4:14:50<1:17:59, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1380/1801 [4:15:01<1:17:47, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1380/1801 [4:15:01<1:17:47, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1381/1801 [4:15:12<1:17:36, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1381/1801 [4:15:12<1:17:36, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1382/1801 [4:15:23<1:17:25, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1382/1801 [4:15:23<1:17:25, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1383/1801 [4:15:34<1:17:14, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1383/1801 [4:15:34<1:17:14, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1384/1801 [4:15:46<1:17:03, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1384/1801 [4:15:46<1:17:03, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1385/1801 [4:15:58<1:16:53, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1385/1801 [4:15:58<1:16:53, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1386/1801 [4:16:09<1:16:41, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1386/1801 [4:16:09<1:16:41, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1387/1801 [4:16:20<1:16:30, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1387/1801 [4:16:20<1:16:30, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1388/1801 [4:16:31<1:16:19, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1388/1801 [4:16:31<1:16:19, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1389/1801 [4:16:42<1:16:08, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1389/1801 [4:16:42<1:16:08, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1390/1801 [4:16:53<1:15:57, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1390/1801 [4:16:53<1:15:57, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1391/1801 [4:17:04<1:15:46, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1391/1801 [4:17:04<1:15:46, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1392/1801 [4:17:17<1:15:35, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1392/1801 [4:17:17<1:15:35, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1393/1801 [4:17:28<1:15:24, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1393/1801 [4:17:28<1:15:24, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1394/1801 [4:17:39<1:15:13, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1394/1801 [4:17:39<1:15:13, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1395/1801 [4:17:50<1:15:02, 0.09it/s, v_num=ufl5] Epoch 0: 77%|███████▋ | 1395/1801 [4:17:50<1:15:02, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1396/1801 [4:18:01<1:14:51, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1396/1801 [4:18:01<1:14:51, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1397/1801 [4:18:13<1:14:40, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1397/1801 [4:18:13<1:14:40, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1398/1801 [4:18:25<1:14:29, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1398/1801 [4:18:25<1:14:29, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1399/1801 [4:18:35<1:14:18, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1399/1801 [4:18:35<1:14:18, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1400/1801 [4:18:48<1:14:07, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1400/1801 [4:18:48<1:14:07, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1401/1801 [4:18:58<1:13:56, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1401/1801 [4:18:58<1:13:56, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1402/1801 [4:19:09<1:13:45, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1402/1801 [4:19:09<1:13:45, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1403/1801 [4:19:20<1:13:34, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1403/1801 [4:19:20<1:13:34, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1404/1801 [4:19:32<1:13:23, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1404/1801 [4:19:32<1:13:23, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1405/1801 [4:19:43<1:13:12, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1405/1801 [4:19:43<1:13:12, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1406/1801 [4:19:54<1:13:01, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1406/1801 [4:19:54<1:13:01, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1407/1801 [4:20:06<1:12:50, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1407/1801 [4:20:06<1:12:50, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1408/1801 [4:20:19<1:12:39, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1408/1801 [4:20:19<1:12:39, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1409/1801 [4:20:29<1:12:28, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1409/1801 [4:20:29<1:12:28, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1410/1801 [4:20:41<1:12:17, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1410/1801 [4:20:41<1:12:17, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1411/1801 [4:20:51<1:12:06, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1411/1801 [4:20:51<1:12:06, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1412/1801 [4:21:02<1:11:54, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1412/1801 [4:21:02<1:11:54, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1413/1801 [4:21:12<1:11:43, 0.09it/s, v_num=ufl5] Epoch 0: 78%|███████▊ | 1413/1801 [4:21:12<1:11:43, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1414/1801 [4:21:24<1:11:32, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1414/1801 [4:21:24<1:11:32, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1415/1801 [4:21:35<1:11:21, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1415/1801 [4:21:35<1:11:21, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1416/1801 [4:21:46<1:11:10, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1416/1801 [4:21:46<1:11:10, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1417/1801 [4:21:58<1:10:59, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1417/1801 [4:21:58<1:10:59, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1418/1801 [4:22:09<1:10:48, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▊ | 1418/1801 [4:22:09<1:10:48, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1419/1801 [4:22:21<1:10:37, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1419/1801 [4:22:21<1:10:37, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1420/1801 [4:22:32<1:10:26, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1420/1801 [4:22:32<1:10:26, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1421/1801 [4:22:43<1:10:15, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1421/1801 [4:22:43<1:10:15, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1422/1801 [4:22:54<1:10:04, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1422/1801 [4:22:54<1:10:04, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1423/1801 [4:23:05<1:09:53, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1423/1801 [4:23:05<1:09:53, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1424/1801 [4:23:18<1:09:42, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1424/1801 [4:23:18<1:09:42, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1425/1801 [4:23:30<1:09:31, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1425/1801 [4:23:30<1:09:31, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1426/1801 [4:23:40<1:09:20, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1426/1801 [4:23:40<1:09:20, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1427/1801 [4:23:51<1:09:09, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1427/1801 [4:23:51<1:09:09, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1428/1801 [4:24:02<1:08:58, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1428/1801 [4:24:02<1:08:58, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1429/1801 [4:24:13<1:08:46, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1429/1801 [4:24:13<1:08:46, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1430/1801 [4:24:24<1:08:35, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1430/1801 [4:24:24<1:08:35, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1431/1801 [4:24:34<1:08:24, 0.09it/s, v_num=ufl5] Epoch 0: 79%|███████▉ | 1431/1801 [4:24:34<1:08:24, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1432/1801 [4:24:46<1:08:13, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1432/1801 [4:24:46<1:08:13, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1433/1801 [4:24:57<1:08:02, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1433/1801 [4:24:57<1:08:02, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1434/1801 [4:25:09<1:07:51, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1434/1801 [4:25:09<1:07:51, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1435/1801 [4:25:20<1:07:40, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1435/1801 [4:25:20<1:07:40, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1436/1801 [4:25:31<1:07:29, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1436/1801 [4:25:31<1:07:29, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1437/1801 [4:25:43<1:07:18, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1437/1801 [4:25:43<1:07:18, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1438/1801 [4:25:54<1:07:07, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1438/1801 [4:25:54<1:07:07, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1439/1801 [4:26:05<1:06:56, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1439/1801 [4:26:05<1:06:56, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1440/1801 [4:26:17<1:06:45, 0.09it/s, v_num=ufl5] Epoch 0: 80%|███████▉ | 1440/1801 [4:26:17<1:06:45, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1441/1801 [4:26:28<1:06:34, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1441/1801 [4:26:28<1:06:34, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1442/1801 [4:26:39<1:06:23, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1442/1801 [4:26:39<1:06:23, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1443/1801 [4:26:50<1:06:12, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1443/1801 [4:26:50<1:06:12, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1444/1801 [4:27:02<1:06:01, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1444/1801 [4:27:02<1:06:01, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1445/1801 [4:27:13<1:05:50, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1445/1801 [4:27:13<1:05:50, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1446/1801 [4:27:24<1:05:38, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1446/1801 [4:27:24<1:05:38, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1447/1801 [4:27:35<1:05:27, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1447/1801 [4:27:35<1:05:27, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1448/1801 [4:27:48<1:05:17, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1448/1801 [4:27:48<1:05:17, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1449/1801 [4:27:59<1:05:06, 0.09it/s, v_num=ufl5] Epoch 0: 80%|████████ | 1449/1801 [4:27:59<1:05:06, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1450/1801 [4:28:09<1:04:54, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1450/1801 [4:28:09<1:04:54, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1451/1801 [4:28:21<1:04:43, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1451/1801 [4:28:21<1:04:43, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1452/1801 [4:28:32<1:04:32, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1452/1801 [4:28:32<1:04:32, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1453/1801 [4:28:44<1:04:21, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1453/1801 [4:28:44<1:04:21, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1454/1801 [4:28:55<1:04:10, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1454/1801 [4:28:55<1:04:10, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1455/1801 [4:29:06<1:03:59, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1455/1801 [4:29:06<1:03:59, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1456/1801 [4:29:18<1:03:48, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1456/1801 [4:29:18<1:03:48, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1457/1801 [4:29:29<1:03:37, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1457/1801 [4:29:29<1:03:37, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1458/1801 [4:29:40<1:03:26, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1458/1801 [4:29:40<1:03:26, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1459/1801 [4:29:51<1:03:15, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1459/1801 [4:29:51<1:03:15, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1460/1801 [4:30:03<1:03:04, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1460/1801 [4:30:03<1:03:04, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1461/1801 [4:30:14<1:02:53, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1461/1801 [4:30:14<1:02:53, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1462/1801 [4:30:25<1:02:42, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1462/1801 [4:30:25<1:02:42, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1463/1801 [4:30:36<1:02:31, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████ | 1463/1801 [4:30:36<1:02:31, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1464/1801 [4:30:49<1:02:20, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1464/1801 [4:30:49<1:02:20, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1465/1801 [4:31:00<1:02:09, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1465/1801 [4:31:00<1:02:09, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1466/1801 [4:31:11<1:01:58, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1466/1801 [4:31:11<1:01:58, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1467/1801 [4:31:22<1:01:47, 0.09it/s, v_num=ufl5] Epoch 0: 81%|████████▏ | 1467/1801 [4:31:22<1:01:47, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1468/1801 [4:31:34<1:01:36, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1468/1801 [4:31:34<1:01:36, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1469/1801 [4:31:44<1:01:24, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1469/1801 [4:31:44<1:01:24, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1470/1801 [4:31:55<1:01:13, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1470/1801 [4:31:55<1:01:13, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1471/1801 [4:32:06<1:01:02, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1471/1801 [4:32:06<1:01:02, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1472/1801 [4:32:18<1:00:51, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1472/1801 [4:32:18<1:00:51, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1473/1801 [4:32:29<1:00:40, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1473/1801 [4:32:29<1:00:40, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1474/1801 [4:32:40<1:00:29, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1474/1801 [4:32:40<1:00:29, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1475/1801 [4:32:52<1:00:18, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1475/1801 [4:32:52<1:00:18, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1476/1801 [4:33:03<1:00:07, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1476/1801 [4:33:03<1:00:07, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1477/1801 [4:33:14<59:56, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1477/1801 [4:33:14<59:56, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1478/1801 [4:33:26<59:45, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1478/1801 [4:33:26<59:45, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1479/1801 [4:33:37<59:34, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1479/1801 [4:33:37<59:34, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1480/1801 [4:33:50<59:23, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1480/1801 [4:33:50<59:23, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1481/1801 [4:34:01<59:12, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1481/1801 [4:34:01<59:12, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1482/1801 [4:34:12<59:01, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1482/1801 [4:34:12<59:01, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1483/1801 [4:34:23<58:50, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1483/1801 [4:34:23<58:50, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1484/1801 [4:34:34<58:39, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1484/1801 [4:34:34<58:39, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1485/1801 [4:34:45<58:28, 0.09it/s, v_num=ufl5] Epoch 0: 82%|████████▏ | 1485/1801 [4:34:45<58:28, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1486/1801 [4:34:57<58:17, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1486/1801 [4:34:57<58:17, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1487/1801 [4:35:08<58:05, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1487/1801 [4:35:08<58:05, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1488/1801 [4:35:20<57:55, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1488/1801 [4:35:20<57:55, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1489/1801 [4:35:31<57:43, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1489/1801 [4:35:31<57:43, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1490/1801 [4:35:42<57:32, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1490/1801 [4:35:42<57:32, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1491/1801 [4:35:53<57:21, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1491/1801 [4:35:53<57:21, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1492/1801 [4:36:04<57:10, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1492/1801 [4:36:04<57:10, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1493/1801 [4:36:15<56:59, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1493/1801 [4:36:15<56:59, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1494/1801 [4:36:26<56:48, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1494/1801 [4:36:26<56:48, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1495/1801 [4:36:38<56:37, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1495/1801 [4:36:38<56:37, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1496/1801 [4:36:49<56:26, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1496/1801 [4:36:49<56:26, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1497/1801 [4:36:59<56:15, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1497/1801 [4:36:59<56:15, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1498/1801 [4:37:11<56:04, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1498/1801 [4:37:11<56:04, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1499/1801 [4:37:22<55:52, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1499/1801 [4:37:22<55:52, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1500/1801 [4:37:33<55:41, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1500/1801 [4:37:33<55:41, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1501/1801 [4:37:45<55:30, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1501/1801 [4:37:45<55:30, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1502/1801 [4:37:56<55:19, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1502/1801 [4:37:56<55:19, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1503/1801 [4:38:07<55:08, 0.09it/s, v_num=ufl5] Epoch 0: 83%|████████▎ | 1503/1801 [4:38:07<55:08, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1504/1801 [4:38:19<54:57, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1504/1801 [4:38:19<54:57, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1505/1801 [4:38:30<54:46, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1505/1801 [4:38:30<54:46, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1506/1801 [4:38:41<54:35, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1506/1801 [4:38:41<54:35, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1507/1801 [4:38:53<54:24, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1507/1801 [4:38:53<54:24, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1508/1801 [4:39:04<54:13, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▎ | 1508/1801 [4:39:04<54:13, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1509/1801 [4:39:15<54:02, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1509/1801 [4:39:15<54:02, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1510/1801 [4:39:27<53:51, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1510/1801 [4:39:27<53:51, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1511/1801 [4:39:38<53:40, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1511/1801 [4:39:38<53:40, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1512/1801 [4:39:50<53:29, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1512/1801 [4:39:50<53:29, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1513/1801 [4:40:02<53:18, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1513/1801 [4:40:02<53:18, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1514/1801 [4:40:13<53:07, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1514/1801 [4:40:13<53:07, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1515/1801 [4:40:25<52:56, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1515/1801 [4:40:25<52:56, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1516/1801 [4:40:36<52:45, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1516/1801 [4:40:36<52:45, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1517/1801 [4:40:48<52:34, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1517/1801 [4:40:48<52:34, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1518/1801 [4:41:00<52:23, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1518/1801 [4:41:00<52:23, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1519/1801 [4:41:11<52:12, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1519/1801 [4:41:11<52:12, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1520/1801 [4:41:24<52:01, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1520/1801 [4:41:24<52:01, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1521/1801 [4:41:36<51:50, 0.09it/s, v_num=ufl5] Epoch 0: 84%|████████▍ | 1521/1801 [4:41:36<51:50, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1522/1801 [4:41:47<51:39, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1522/1801 [4:41:47<51:39, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1523/1801 [4:41:58<51:28, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1523/1801 [4:41:58<51:28, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1524/1801 [4:42:09<51:17, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1524/1801 [4:42:09<51:17, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1525/1801 [4:42:20<51:05, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1525/1801 [4:42:20<51:05, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1526/1801 [4:42:31<50:54, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1526/1801 [4:42:31<50:54, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1527/1801 [4:42:43<50:43, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1527/1801 [4:42:43<50:43, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1528/1801 [4:42:55<50:32, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1528/1801 [4:42:55<50:32, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1529/1801 [4:43:06<50:21, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1529/1801 [4:43:06<50:21, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1530/1801 [4:43:17<50:10, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▍ | 1530/1801 [4:43:17<50:10, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1531/1801 [4:43:29<49:59, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1531/1801 [4:43:29<49:59, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1532/1801 [4:43:40<49:48, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1532/1801 [4:43:40<49:48, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1533/1801 [4:43:51<49:37, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1533/1801 [4:43:51<49:37, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1534/1801 [4:44:02<49:26, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1534/1801 [4:44:02<49:26, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1535/1801 [4:44:13<49:15, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1535/1801 [4:44:13<49:15, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1536/1801 [4:44:25<49:04, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1536/1801 [4:44:25<49:04, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1537/1801 [4:44:37<48:53, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1537/1801 [4:44:37<48:53, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1538/1801 [4:44:48<48:42, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1538/1801 [4:44:48<48:42, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1539/1801 [4:44:59<48:30, 0.09it/s, v_num=ufl5] Epoch 0: 85%|████████▌ | 1539/1801 [4:44:59<48:30, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1540/1801 [4:45:09<48:19, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1540/1801 [4:45:09<48:19, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1541/1801 [4:45:20<48:08, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1541/1801 [4:45:20<48:08, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1542/1801 [4:45:31<47:57, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1542/1801 [4:45:31<47:57, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1543/1801 [4:45:42<47:46, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1543/1801 [4:45:42<47:46, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1544/1801 [4:45:54<47:35, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1544/1801 [4:45:54<47:35, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1545/1801 [4:46:06<47:24, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1545/1801 [4:46:06<47:24, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1546/1801 [4:46:17<47:13, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1546/1801 [4:46:17<47:13, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1547/1801 [4:46:28<47:02, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1547/1801 [4:46:28<47:02, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1548/1801 [4:46:38<46:50, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1548/1801 [4:46:38<46:50, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1549/1801 [4:46:49<46:39, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1549/1801 [4:46:49<46:39, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1550/1801 [4:47:00<46:28, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1550/1801 [4:47:00<46:28, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1551/1801 [4:47:11<46:17, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1551/1801 [4:47:11<46:17, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1552/1801 [4:47:23<46:06, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1552/1801 [4:47:23<46:06, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1553/1801 [4:47:35<45:55, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▌ | 1553/1801 [4:47:35<45:55, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1554/1801 [4:47:45<45:44, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1554/1801 [4:47:45<45:44, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1555/1801 [4:47:56<45:33, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1555/1801 [4:47:56<45:33, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1556/1801 [4:48:08<45:22, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1556/1801 [4:48:08<45:22, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1557/1801 [4:48:18<45:10, 0.09it/s, v_num=ufl5] Epoch 0: 86%|████████▋ | 1557/1801 [4:48:18<45:10, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1558/1801 [4:48:29<44:59, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1558/1801 [4:48:29<44:59, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1559/1801 [4:48:40<44:48, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1559/1801 [4:48:40<44:48, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1560/1801 [4:48:52<44:37, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1560/1801 [4:48:52<44:37, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1561/1801 [4:49:03<44:26, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1561/1801 [4:49:03<44:26, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1562/1801 [4:49:14<44:15, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1562/1801 [4:49:14<44:15, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1563/1801 [4:49:24<44:04, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1563/1801 [4:49:24<44:04, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1564/1801 [4:49:35<43:53, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1564/1801 [4:49:35<43:53, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1565/1801 [4:49:47<43:41, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1565/1801 [4:49:47<43:41, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1566/1801 [4:49:58<43:30, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1566/1801 [4:49:58<43:30, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1567/1801 [4:50:09<43:19, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1567/1801 [4:50:09<43:19, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1568/1801 [4:50:21<43:08, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1568/1801 [4:50:21<43:08, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1569/1801 [4:50:31<42:57, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1569/1801 [4:50:31<42:57, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1570/1801 [4:50:42<42:46, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1570/1801 [4:50:42<42:46, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1571/1801 [4:50:54<42:35, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1571/1801 [4:50:54<42:35, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1572/1801 [4:51:04<42:24, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1572/1801 [4:51:04<42:24, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1573/1801 [4:51:16<42:13, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1573/1801 [4:51:16<42:13, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1574/1801 [4:51:27<42:02, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1574/1801 [4:51:27<42:02, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1575/1801 [4:51:38<41:50, 0.09it/s, v_num=ufl5] Epoch 0: 87%|████████▋ | 1575/1801 [4:51:38<41:50, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1576/1801 [4:51:50<41:39, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1576/1801 [4:51:50<41:39, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1577/1801 [4:52:00<41:28, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1577/1801 [4:52:00<41:28, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1578/1801 [4:52:11<41:17, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1578/1801 [4:52:11<41:17, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1579/1801 [4:52:23<41:06, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1579/1801 [4:52:23<41:06, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1580/1801 [4:52:34<40:55, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1580/1801 [4:52:34<40:55, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1581/1801 [4:52:45<40:44, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1581/1801 [4:52:45<40:44, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1582/1801 [4:52:56<40:33, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1582/1801 [4:52:56<40:33, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1583/1801 [4:53:07<40:22, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1583/1801 [4:53:07<40:22, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1584/1801 [4:53:19<40:11, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1584/1801 [4:53:19<40:11, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1585/1801 [4:53:30<39:59, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1585/1801 [4:53:30<39:59, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1586/1801 [4:53:40<39:48, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1586/1801 [4:53:40<39:48, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1587/1801 [4:53:52<39:37, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1587/1801 [4:53:52<39:37, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1588/1801 [4:54:03<39:26, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1588/1801 [4:54:03<39:26, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1589/1801 [4:54:14<39:15, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1589/1801 [4:54:14<39:15, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1590/1801 [4:54:25<39:04, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1590/1801 [4:54:25<39:04, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1591/1801 [4:54:36<38:53, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1591/1801 [4:54:36<38:53, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1592/1801 [4:54:48<38:42, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1592/1801 [4:54:48<38:42, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1593/1801 [4:54:59<38:31, 0.09it/s, v_num=ufl5] Epoch 0: 88%|████████▊ | 1593/1801 [4:54:59<38:31, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1594/1801 [4:55:10<38:19, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1594/1801 [4:55:10<38:19, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1595/1801 [4:55:20<38:08, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1595/1801 [4:55:20<38:08, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1596/1801 [4:55:32<37:57, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1596/1801 [4:55:32<37:57, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1597/1801 [4:55:42<37:46, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1597/1801 [4:55:42<37:46, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1598/1801 [4:55:54<37:35, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▊ | 1598/1801 [4:55:54<37:35, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1599/1801 [4:56:04<37:24, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1599/1801 [4:56:04<37:24, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1600/1801 [4:56:16<37:13, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1600/1801 [4:56:16<37:13, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1601/1801 [4:56:27<37:02, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1601/1801 [4:56:27<37:02, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1602/1801 [4:56:39<36:51, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1602/1801 [4:56:39<36:51, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1603/1801 [4:56:49<36:39, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1603/1801 [4:56:49<36:39, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1604/1801 [4:57:00<36:28, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1604/1801 [4:57:00<36:28, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1605/1801 [4:57:11<36:17, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1605/1801 [4:57:11<36:17, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1606/1801 [4:57:22<36:06, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1606/1801 [4:57:22<36:06, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1607/1801 [4:57:33<35:55, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1607/1801 [4:57:33<35:55, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1608/1801 [4:57:45<35:44, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1608/1801 [4:57:45<35:44, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1609/1801 [4:57:56<35:33, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1609/1801 [4:57:56<35:33, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1610/1801 [4:58:07<35:22, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1610/1801 [4:58:07<35:22, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1611/1801 [4:58:18<35:10, 0.09it/s, v_num=ufl5] Epoch 0: 89%|████████▉ | 1611/1801 [4:58:18<35:10, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1612/1801 [4:58:30<34:59, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1612/1801 [4:58:30<34:59, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1613/1801 [4:58:40<34:48, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1613/1801 [4:58:40<34:48, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1614/1801 [4:58:51<34:37, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1614/1801 [4:58:51<34:37, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1615/1801 [4:59:03<34:26, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1615/1801 [4:59:03<34:26, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1616/1801 [4:59:15<34:15, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1616/1801 [4:59:15<34:15, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1617/1801 [4:59:26<34:04, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1617/1801 [4:59:26<34:04, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1618/1801 [4:59:37<33:53, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1618/1801 [4:59:37<33:53, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1619/1801 [4:59:47<33:42, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1619/1801 [4:59:47<33:42, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1620/1801 [4:59:58<33:30, 0.09it/s, v_num=ufl5] Epoch 0: 90%|████████▉ | 1620/1801 [4:59:58<33:30, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1621/1801 [5:00:09<33:19, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1621/1801 [5:00:09<33:19, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1622/1801 [5:00:20<33:08, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1622/1801 [5:00:20<33:08, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1623/1801 [5:00:31<32:57, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1623/1801 [5:00:31<32:57, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1624/1801 [5:00:44<32:46, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1624/1801 [5:00:44<32:46, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1625/1801 [5:00:55<32:35, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1625/1801 [5:00:55<32:35, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1626/1801 [5:01:06<32:24, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1626/1801 [5:01:06<32:24, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1627/1801 [5:01:18<32:13, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1627/1801 [5:01:18<32:13, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1628/1801 [5:01:29<32:02, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1628/1801 [5:01:29<32:02, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1629/1801 [5:01:41<31:51, 0.09it/s, v_num=ufl5] Epoch 0: 90%|█████████ | 1629/1801 [5:01:41<31:51, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1630/1801 [5:01:52<31:40, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1630/1801 [5:01:52<31:40, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1631/1801 [5:02:03<31:29, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1631/1801 [5:02:03<31:29, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1632/1801 [5:02:16<31:18, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1632/1801 [5:02:16<31:18, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1633/1801 [5:02:27<31:06, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1633/1801 [5:02:27<31:06, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1634/1801 [5:02:38<30:55, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1634/1801 [5:02:38<30:55, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1635/1801 [5:02:49<30:44, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1635/1801 [5:02:49<30:44, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1636/1801 [5:03:00<30:33, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1636/1801 [5:03:00<30:33, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1637/1801 [5:03:10<30:22, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1637/1801 [5:03:10<30:22, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1638/1801 [5:03:22<30:11, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1638/1801 [5:03:22<30:11, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1639/1801 [5:03:33<30:00, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1639/1801 [5:03:33<30:00, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1640/1801 [5:03:45<29:49, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1640/1801 [5:03:45<29:49, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1641/1801 [5:03:57<29:38, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1641/1801 [5:03:57<29:38, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1642/1801 [5:04:08<29:27, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1642/1801 [5:04:08<29:27, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1643/1801 [5:04:19<29:15, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████ | 1643/1801 [5:04:19<29:15, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1644/1801 [5:04:30<29:04, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1644/1801 [5:04:30<29:04, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1645/1801 [5:04:40<28:53, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1645/1801 [5:04:40<28:53, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1646/1801 [5:04:51<28:42, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1646/1801 [5:04:51<28:42, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1647/1801 [5:05:02<28:31, 0.09it/s, v_num=ufl5] Epoch 0: 91%|█████████▏| 1647/1801 [5:05:02<28:31, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1648/1801 [5:05:14<28:20, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1648/1801 [5:05:14<28:20, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1649/1801 [5:05:25<28:09, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1649/1801 [5:05:25<28:09, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1650/1801 [5:05:36<27:58, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1650/1801 [5:05:36<27:58, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1651/1801 [5:05:47<27:46, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1651/1801 [5:05:47<27:46, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1652/1801 [5:05:59<27:35, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1652/1801 [5:05:59<27:35, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1653/1801 [5:06:10<27:24, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1653/1801 [5:06:10<27:24, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1654/1801 [5:06:21<27:13, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1654/1801 [5:06:21<27:13, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1655/1801 [5:06:32<27:02, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1655/1801 [5:06:32<27:02, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1656/1801 [5:06:44<26:51, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1656/1801 [5:06:44<26:51, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1657/1801 [5:06:55<26:40, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1657/1801 [5:06:55<26:40, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1658/1801 [5:07:06<26:29, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1658/1801 [5:07:06<26:29, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1659/1801 [5:07:17<26:18, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1659/1801 [5:07:17<26:18, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1660/1801 [5:07:27<26:06, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1660/1801 [5:07:27<26:06, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1661/1801 [5:07:38<25:55, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1661/1801 [5:07:38<25:55, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1662/1801 [5:07:49<25:44, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1662/1801 [5:07:49<25:44, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1663/1801 [5:08:01<25:33, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1663/1801 [5:08:01<25:33, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1664/1801 [5:08:13<25:22, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1664/1801 [5:08:13<25:22, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1665/1801 [5:08:24<25:11, 0.09it/s, v_num=ufl5] Epoch 0: 92%|█████████▏| 1665/1801 [5:08:24<25:11, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1666/1801 [5:08:35<25:00, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1666/1801 [5:08:35<25:00, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1667/1801 [5:08:46<24:49, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1667/1801 [5:08:46<24:49, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1668/1801 [5:08:58<24:38, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1668/1801 [5:08:58<24:38, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1669/1801 [5:09:09<24:27, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1669/1801 [5:09:09<24:27, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1670/1801 [5:09:19<24:15, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1670/1801 [5:09:19<24:15, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1671/1801 [5:09:30<24:04, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1671/1801 [5:09:30<24:04, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1672/1801 [5:09:41<23:53, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1672/1801 [5:09:41<23:53, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1673/1801 [5:09:53<23:42, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1673/1801 [5:09:53<23:42, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1674/1801 [5:10:04<23:31, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1674/1801 [5:10:04<23:31, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1675/1801 [5:10:16<23:20, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1675/1801 [5:10:16<23:20, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1676/1801 [5:10:27<23:09, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1676/1801 [5:10:27<23:09, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1677/1801 [5:10:38<22:58, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1677/1801 [5:10:38<22:58, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1678/1801 [5:10:49<22:47, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1678/1801 [5:10:49<22:47, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1679/1801 [5:10:59<22:35, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1679/1801 [5:10:59<22:35, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1680/1801 [5:11:11<22:24, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1680/1801 [5:11:11<22:24, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1681/1801 [5:11:22<22:13, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1681/1801 [5:11:22<22:13, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1682/1801 [5:11:33<22:02, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1682/1801 [5:11:33<22:02, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1683/1801 [5:11:44<21:51, 0.09it/s, v_num=ufl5] Epoch 0: 93%|█████████▎| 1683/1801 [5:11:44<21:51, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1684/1801 [5:11:55<21:40, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1684/1801 [5:11:55<21:40, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1685/1801 [5:12:06<21:29, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1685/1801 [5:12:06<21:29, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1686/1801 [5:12:17<21:18, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1686/1801 [5:12:17<21:18, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1687/1801 [5:12:28<21:06, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1687/1801 [5:12:28<21:06, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1688/1801 [5:12:40<20:55, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▎| 1688/1801 [5:12:40<20:55, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1689/1801 [5:12:52<20:44, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1689/1801 [5:12:52<20:44, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1690/1801 [5:13:02<20:33, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1690/1801 [5:13:02<20:33, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1691/1801 [5:13:14<20:22, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1691/1801 [5:13:14<20:22, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1692/1801 [5:13:25<20:11, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1692/1801 [5:13:25<20:11, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1693/1801 [5:13:36<20:00, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1693/1801 [5:13:36<20:00, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1694/1801 [5:13:47<19:49, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1694/1801 [5:13:47<19:49, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1695/1801 [5:13:58<19:38, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1695/1801 [5:13:58<19:38, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1696/1801 [5:14:09<19:27, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1696/1801 [5:14:09<19:27, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1697/1801 [5:14:20<19:15, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1697/1801 [5:14:20<19:15, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1698/1801 [5:14:31<19:04, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1698/1801 [5:14:31<19:04, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1699/1801 [5:14:42<18:53, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1699/1801 [5:14:42<18:53, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1700/1801 [5:14:53<18:42, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1700/1801 [5:14:53<18:42, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1701/1801 [5:15:04<18:31, 0.09it/s, v_num=ufl5] Epoch 0: 94%|█████████▍| 1701/1801 [5:15:04<18:31, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1702/1801 [5:15:15<18:20, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1702/1801 [5:15:15<18:20, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1703/1801 [5:15:26<18:09, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1703/1801 [5:15:26<18:09, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1704/1801 [5:15:39<17:58, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1704/1801 [5:15:39<17:58, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1705/1801 [5:15:50<17:46, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1705/1801 [5:15:50<17:46, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1706/1801 [5:16:00<17:35, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1706/1801 [5:16:00<17:35, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1707/1801 [5:16:11<17:24, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1707/1801 [5:16:11<17:24, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1708/1801 [5:16:22<17:13, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1708/1801 [5:16:22<17:13, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1709/1801 [5:16:33<17:02, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1709/1801 [5:16:33<17:02, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1710/1801 [5:16:44<16:51, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▍| 1710/1801 [5:16:44<16:51, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1711/1801 [5:16:55<16:40, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1711/1801 [5:16:55<16:40, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1712/1801 [5:17:07<16:29, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1712/1801 [5:17:07<16:29, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1713/1801 [5:17:18<16:18, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1713/1801 [5:17:18<16:18, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1714/1801 [5:17:29<16:06, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1714/1801 [5:17:29<16:06, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1715/1801 [5:17:40<15:55, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1715/1801 [5:17:40<15:55, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1716/1801 [5:17:51<15:44, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1716/1801 [5:17:51<15:44, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1717/1801 [5:18:02<15:33, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1717/1801 [5:18:02<15:33, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1718/1801 [5:18:13<15:22, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1718/1801 [5:18:13<15:22, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1719/1801 [5:18:23<15:11, 0.09it/s, v_num=ufl5] Epoch 0: 95%|█████████▌| 1719/1801 [5:18:23<15:11, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1720/1801 [5:18:35<15:00, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1720/1801 [5:18:35<15:00, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1721/1801 [5:18:45<14:49, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1721/1801 [5:18:45<14:49, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1722/1801 [5:18:57<14:37, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1722/1801 [5:18:57<14:37, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1723/1801 [5:19:08<14:26, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1723/1801 [5:19:08<14:26, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1724/1801 [5:19:19<14:15, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1724/1801 [5:19:19<14:15, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1725/1801 [5:19:31<14:04, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1725/1801 [5:19:31<14:04, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1726/1801 [5:19:42<13:53, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1726/1801 [5:19:42<13:53, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1727/1801 [5:19:54<13:42, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1727/1801 [5:19:54<13:42, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1728/1801 [5:20:07<13:31, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1728/1801 [5:20:07<13:31, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1729/1801 [5:20:18<13:20, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1729/1801 [5:20:18<13:20, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1730/1801 [5:20:30<13:09, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1730/1801 [5:20:30<13:09, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1731/1801 [5:20:41<12:58, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1731/1801 [5:20:41<12:58, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1732/1801 [5:20:53<12:47, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1732/1801 [5:20:53<12:47, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1733/1801 [5:21:05<12:35, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▌| 1733/1801 [5:21:05<12:35, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1734/1801 [5:21:17<12:24, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1734/1801 [5:21:17<12:24, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1735/1801 [5:21:28<12:13, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1735/1801 [5:21:28<12:13, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1736/1801 [5:21:40<12:02, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1736/1801 [5:21:40<12:02, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1737/1801 [5:21:52<11:51, 0.09it/s, v_num=ufl5] Epoch 0: 96%|█████████▋| 1737/1801 [5:21:52<11:51, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1738/1801 [5:22:03<11:40, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1738/1801 [5:22:03<11:40, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1739/1801 [5:22:15<11:29, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1739/1801 [5:22:15<11:29, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1740/1801 [5:22:27<11:18, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1740/1801 [5:22:27<11:18, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1741/1801 [5:22:39<11:07, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1741/1801 [5:22:39<11:07, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1742/1801 [5:22:50<10:56, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1742/1801 [5:22:50<10:56, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1743/1801 [5:23:02<10:44, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1743/1801 [5:23:02<10:44, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1744/1801 [5:23:15<10:33, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1744/1801 [5:23:15<10:33, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1745/1801 [5:23:26<10:22, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1745/1801 [5:23:26<10:22, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1746/1801 [5:23:37<10:11, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1746/1801 [5:23:37<10:11, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1747/1801 [5:23:47<10:00, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1747/1801 [5:23:47<10:00, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1748/1801 [5:23:59<09:49, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1748/1801 [5:23:59<09:49, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1749/1801 [5:24:10<09:38, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1749/1801 [5:24:10<09:38, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1750/1801 [5:24:21<09:27, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1750/1801 [5:24:21<09:27, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1751/1801 [5:24:32<09:16, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1751/1801 [5:24:32<09:16, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1752/1801 [5:24:43<09:04, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1752/1801 [5:24:43<09:04, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1753/1801 [5:24:55<08:53, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1753/1801 [5:24:55<08:53, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1754/1801 [5:25:06<08:42, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1754/1801 [5:25:06<08:42, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1755/1801 [5:25:17<08:31, 0.09it/s, v_num=ufl5] Epoch 0: 97%|█████████▋| 1755/1801 [5:25:17<08:31, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1756/1801 [5:25:28<08:20, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1756/1801 [5:25:28<08:20, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1757/1801 [5:25:39<08:09, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1757/1801 [5:25:39<08:09, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1758/1801 [5:25:50<07:58, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1758/1801 [5:25:50<07:58, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1759/1801 [5:26:01<07:47, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1759/1801 [5:26:01<07:47, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1760/1801 [5:26:14<07:35, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1760/1801 [5:26:14<07:35, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1761/1801 [5:26:25<07:24, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1761/1801 [5:26:25<07:24, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1762/1801 [5:26:35<07:13, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1762/1801 [5:26:35<07:13, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1763/1801 [5:26:47<07:02, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1763/1801 [5:26:47<07:02, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1764/1801 [5:26:58<06:51, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1764/1801 [5:26:58<06:51, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1765/1801 [5:27:10<06:40, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1765/1801 [5:27:10<06:40, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1766/1801 [5:27:20<06:29, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1766/1801 [5:27:20<06:29, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1767/1801 [5:27:32<06:18, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1767/1801 [5:27:32<06:18, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1768/1801 [5:27:45<06:07, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1768/1801 [5:27:45<06:07, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1769/1801 [5:27:56<05:55, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1769/1801 [5:27:56<05:55, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1770/1801 [5:28:07<05:44, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1770/1801 [5:28:07<05:44, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1771/1801 [5:28:18<05:33, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1771/1801 [5:28:18<05:33, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1772/1801 [5:28:30<05:22, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1772/1801 [5:28:30<05:22, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1773/1801 [5:28:40<05:11, 0.09it/s, v_num=ufl5] Epoch 0: 98%|█████████▊| 1773/1801 [5:28:40<05:11, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1774/1801 [5:28:52<05:00, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1774/1801 [5:28:52<05:00, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1775/1801 [5:29:03<04:49, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1775/1801 [5:29:03<04:49, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1776/1801 [5:29:15<04:38, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1776/1801 [5:29:15<04:38, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1777/1801 [5:29:26<04:26, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1777/1801 [5:29:26<04:26, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1778/1801 [5:29:37<04:15, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▊| 1778/1801 [5:29:37<04:15, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1779/1801 [5:29:49<04:04, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1779/1801 [5:29:49<04:04, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1780/1801 [5:30:01<03:53, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1780/1801 [5:30:01<03:53, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1781/1801 [5:30:12<03:42, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1781/1801 [5:30:12<03:42, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1782/1801 [5:30:22<03:31, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1782/1801 [5:30:22<03:31, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1783/1801 [5:30:32<03:20, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1783/1801 [5:30:32<03:20, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1784/1801 [5:30:44<03:09, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1784/1801 [5:30:44<03:09, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1785/1801 [5:30:56<02:57, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1785/1801 [5:30:56<02:57, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1786/1801 [5:31:07<02:46, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1786/1801 [5:31:07<02:46, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1787/1801 [5:31:19<02:35, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1787/1801 [5:31:19<02:35, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1788/1801 [5:31:29<02:24, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1788/1801 [5:31:29<02:24, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1789/1801 [5:31:40<02:13, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1789/1801 [5:31:40<02:13, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1790/1801 [5:31:51<02:02, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1790/1801 [5:31:51<02:02, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1791/1801 [5:32:03<01:51, 0.09it/s, v_num=ufl5] Epoch 0: 99%|█████████▉| 1791/1801 [5:32:03<01:51, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1792/1801 [5:32:15<01:40, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1792/1801 [5:32:15<01:40, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1793/1801 [5:32:25<01:28, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1793/1801 [5:32:25<01:28, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1794/1801 [5:32:36<01:17, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1794/1801 [5:32:36<01:17, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1795/1801 [5:32:48<01:06, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1795/1801 [5:32:48<01:06, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1796/1801 [5:32:59<00:55, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1796/1801 [5:32:59<00:55, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1797/1801 [5:33:09<00:44, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1797/1801 [5:33:09<00:44, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1798/1801 [5:33:21<00:33, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1798/1801 [5:33:21<00:33, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1799/1801 [5:33:32<00:22, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1799/1801 [5:33:32<00:22, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1800/1801 [5:33:44<00:11, 0.09it/s, v_num=ufl5] Epoch 0: 100%|█████████▉| 1800/1801 [5:33:44<00:11, 0.09it/s, v_num=ufl5] Epoch 0: 100%|██████████| 1801/1801 [5:33:55<00:00, 0.09it/s, v_num=ufl5] Epoch 0: 100%|██████████| 1801/1801 [5:33:55<00:00, 0.09it/s, v_num=ufl5] Epoch 0: 100%|██████████| 1801/1801 [5:33:55<00:00, 0.09it/s, v_num=ufl5]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_word +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/serialization.py", line 967, in save + _save( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/serialization.py", line 1268, in _save + zip_file.write_record(name, storage, num_bytes) +RuntimeError: [enforce fail at inline_container.cc:858] . PytorchStreamWriter failed writing file data/1051: file write failed + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): +[rank2]: Traceback (most recent call last): +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/serialization.py", line 967, in save +[rank2]: _save( +[wandb: +wandb: 🚀 View run Test-AC-critic_expectile_0.9-inv_temp_1.0 at:  +W0915 22:45:16.312028 1515135 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 1515467 closing signal SIGTERM +W0915 22:45:16.314096 1515135 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 1515469 closing signal SIGTERM +W0915 22:45:16.315508 1515135 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 1515473 closing signal SIGTERM +E0915 22:45:16.781522 1515135 site-packages/torch/distributed/elastic/multiprocessing/api.py:874] failed (exitcode: 1) local_rank: 2 (pid: 1515472) of binary: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/bin/accelerate", line 7, in + sys.exit(main()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main + args.func(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/launch.py", line 1220, in launch_command + deepspeed_launcher(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/launch.py", line 906, in deepspeed_launcher + distrib_run.run(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/run.py", line 892, in run + elastic_launch( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 143, in __call__ + return launch_agent(self._config, self._entrypoint, list(args)) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 277, in launch_agent + raise ChildFailedError( +torch.distributed.elastic.multiprocessing.errors.ChildFailedError: +============================================================ +/home/jiashuo/codes/OfflineArcher/main.py FAILED +------------------------------------------------------------ +Failures: + +------------------------------------------------------------ +Root Cause (first observed failure): +[0]: + time : 2025-09-15_22:45:16 + host : somea6k.local + rank : 2 (local_rank: 2) + exitcode : 1 (pid: 1515472) + error_file: + traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html +============================================================ diff --git a/Qwen3-14B_Strategic_log.txt b/Qwen3-14B_Strategic_log.txt new file mode 100644 index 0000000000000000000000000000000000000000..09879199ece3d8017bafe85b86c86d3cebe2fb62 --- /dev/null +++ b/Qwen3-14B_Strategic_log.txt @@ -0,0 +1,12 @@ +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/bin/accelerate", line 7, in + sys.exit(main()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main + args.func(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/launch.py", line 1213, in launch_command + args, defaults, mp_from_config_flag = _validate_launch_command(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/launch.py", line 1043, in _validate_launch_command + defaults = load_config_from_file(args.config_file) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/config/config_args.py", line 46, in load_config_from_file + raise FileNotFoundError( +FileNotFoundError: The passed configuration file `/home/jiashuo/codes/OfflineArcher/deepspeed_zero3.syaml` does not exist. Please pass an existing file to `accelerate launch`, or use the default one created through `accelerate config` and run `accelerate launch` without the `--config_file` argument. diff --git a/Qwen3-14B_Word_log.txt b/Qwen3-14B_Word_log.txt new file mode 100644 index 0000000000000000000000000000000000000000..723cdeee6c20d8a418d0277adef5bce6991f9cc1 --- /dev/null +++ b/Qwen3-14B_Word_log.txt @@ -0,0 +1,285 @@ +The following values were not passed to `accelerate launch` and had defaults used instead: + `--num_cpu_threads_per_process` was set to `16` to improve out-of-box performance when training on CPUs +To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`. +[2025-09-16 23:00:41,059] [INFO] [real_accelerator.py:260:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-09-16 23:00:42,764] [INFO] [logging.py:107:log_dist] [Rank -1] [TorchCheckpointEngine] Initialized with serialization = False +W0916 23:00:42.933299 185693 site-packages/torch/distributed/run.py:774] +W0916 23:00:42.933299 185693 site-packages/torch/distributed/run.py:774] ***************************************** +W0916 23:00:42.933299 185693 site-packages/torch/distributed/run.py:774] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed. +W0916 23:00:42.933299 185693 site-packages/torch/distributed/run.py:774] ***************************************** +[rank: 0] Seed set to 42 +[rank: 3] Seed set to 42 +[rank: 2] Seed set to 42 +[rank: 1] Seed set to 42 +[rank: 3] Seed set to 42 +[rank: 0] Seed set to 42 +`torch_dtype` is deprecated! Use `dtype` instead! +[rank: 2] Seed set to 42 + Loading checkpoint shards: 0%| | 0/6 [00:00 +[rank2]: cli_main() +[rank2]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank2]: cli = LightningCLI( +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank2]: self._run_subcommand(self.subcommand) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank2]: fn(**fn_kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank2]: call._call_and_handle_interrupt( +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank2]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank2]: return function(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank2]: self._run(model, ckpt_path=ckpt_path) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank2]: results = self._run_stage() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank2]: self.fit_loop.run() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank2]: self.advance() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank2]: self.epoch_loop.run(self._data_fetcher) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank2]: self.advance(data_fetcher) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank2]: batch_output = self.manual_optimization.run(kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank2]: self.advance(kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank2]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank2]: output = fn(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank2]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank2]: wrapper_output = wrapper_module(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank2]: return forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank2]: ret_val = func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank2]: loss = self.module(*inputs, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank2]: result = forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank2]: out = method(*_args, **_kwargs) +[rank2]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 533, in training_step +[rank2]: actor_loss, actor_log = self.actor_loss(batch) +[rank2]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 611, in +[rank2]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank2]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 615, in awr_loss +[rank2]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank2]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 130, in get_logsum_prob +[rank2]: generated_probabilities = self.to_tokens_and_logprobs(alltext) +[rank2]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 163, in to_tokens_and_logprobs +[rank2]: outputs = self.model(input_ids) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank2]: result = forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward +[rank2]: return self.base_model( +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank2]: result = forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward +[rank2]: return self.model.forward(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 940, in wrapper +[rank2]: output = func(self, *args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward +[rank2]: outputs: BaseModelOutputWithPast = self.model( +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank2]: result = forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 1064, in wrapper +[rank2]: outputs = func(self, *args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 410, in forward +[rank2]: hidden_states = decoder_layer( +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ +[rank2]: return super().__call__(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank2]: result = forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func +[rank2]: return func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 275, in forward +[rank2]: hidden_states = self.mlp(hidden_states) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank2]: result = forward_call(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 82, in forward +[rank2]: down_proj = self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank2]: return self._call_impl(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank2]: return inner() +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1816, in inner +[rank2]: args_result = hook(self, args) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 929, in _fn +[rank2]: return fn(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/parameter_offload.py", line 298, in _pre_forward_module_hook +[rank2]: self.pre_sub_module_forward_function(module) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context +[rank2]: return func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/parameter_offload.py", line 473, in pre_sub_module_forward_function +[rank2]: param_coordinator.fetch_sub_module(sub_module, forward=True) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 929, in _fn +[rank2]: return fn(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank2]: ret_val = func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context +[rank2]: return func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partitioned_param_coordinator.py", line 415, in fetch_sub_module +[rank2]: self.__all_gather_params(params_to_prefetch, forward) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank2]: ret_val = func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partitioned_param_coordinator.py", line 475, in __all_gather_params +[rank2]: self.__all_gather_params_(nonquantized_params, forward, quantize=self.zero_quantized_weights) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partitioned_param_coordinator.py", line 504, in __all_gather_params_ +[rank2]: handle = param_group[0].all_gather_coalesced(param_group, quantize=quantize) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank2]: ret_val = func(*args, **kwargs) +[rank2]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partition_parameters.py", line 1325, in all_gather_coalesced +[rank2]: param_buffer = torch.empty( +[rank2]: torch.AcceleratorError: CUDA error: out of memory +[rank2]: Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. + +W0916 23:03:30.410014 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 185929 closing signal SIGTERM +W0916 23:03:30.412096 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 185930 closing signal SIGTERM +W0916 23:03:30.414456 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 185932 closing signal SIGTERM +W0916 23:04:00.415664 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:919] Unable to shutdown process 185929 via Signals.SIGTERM, forcefully exiting via Signals.SIGKILL +W0916 23:04:00.757989 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:919] Unable to shutdown process 185930 via Signals.SIGTERM, forcefully exiting via Signals.SIGKILL +W0916 23:04:01.076958 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:919] Unable to shutdown process 185932 via Signals.SIGTERM, forcefully exiting via Signals.SIGKILL +E0916 23:04:01.370846 185693 site-packages/torch/distributed/elastic/multiprocessing/api.py:874] failed (exitcode: 1) local_rank: 2 (pid: 185931) of binary: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/bin/accelerate", line 7, in + sys.exit(main()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/accelerate_cli.py", line 50, in main + args.func(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/launch.py", line 1220, in launch_command + deepspeed_launcher(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/accelerate/commands/launch.py", line 906, in deepspeed_launcher + distrib_run.run(args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/run.py", line 892, in run + elastic_launch( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 143, in __call__ + return launch_agent(self._config, self._entrypoint, list(args)) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/launcher/api.py", line 277, in launch_agent + raise ChildFailedError( +torch.distributed.elastic.multiprocessing.errors.ChildFailedError: +============================================================ +/home/jiashuo/codes/OfflineArcher/main.py FAILED +------------------------------------------------------------ +Failures: + +------------------------------------------------------------ +Root Cause (first observed failure): +[0]: + time : 2025-09-16_23:03:30 + host : somea6k.local + rank : 2 (local_rank: 2) + exitcode : 1 (pid: 185931) + error_file: + traceback : To enable traceback see: https://pytorch.org/docs/stable/elastic/errors.html +============================================================ diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..37db70c85bb1e6061f49e796bad6af2ecacd69a3 --- /dev/null +++ b/README.md @@ -0,0 +1,57 @@ +# OfflineArcher +Research Code for the Offline Experiments of "ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL" + +[Yifei Zhou](https://yifeizhou02.github.io/), [Andrea Zanette](https://azanette.com/), [Jiayi Pan](https://www.jiayipan.me/), [Aviral Kumar](https://aviralkumar2907.github.io/), [Sergey Levine](https://people.eecs.berkeley.edu/~svlevine/) + +![archer_diagram 001](https://github.com/YifeiZhou02/ArCHer/assets/83000332/b874432a-d330-49a5-906c-bba37e17f831) + + +This repo supports the following methods: + +- [Offline ArCHer][1] +- Offline Filtered BC +- Offline BC + +[1]: https://github.com/YifeiZhou02/ArCHer + +And the following environments +- [Twenty Questions][2] + +[2]: https://lmrl-gym.github.io/ + + +## Quick Start +### 1. Install Dependencies +```bash +conda create -n archer python==3.10 +conda activate archer + +git clone https://github.com/andreazanette/OfflineArcher +cd OfflineArcher +python -m pip install -e . +``` +### 2. Download Datasets and Oracles +Offline datasets and Oracles checkpoints used in the paper can be found [here](https://drive.google.com/drive/folders/1pRocQI0Jv479G4vNMtQn1JOq8Shf2B6U?usp=sharing). +You will need to create an "oracles" and "datasets" folder and put the oracle and dataset in such folders. +The oracle for Twenty Questions should be named 20q_t5_oracle.pt and the dataset should be called "twenty_questions.json". + +### 3. Run Experiments +You can directly run experiments by runnig the launch scripts. For example, in order to lauch Offline Archer on Twenty Question simply run +```bash +. submit_OfflineArcher_TwentyQuestions.sh +``` +The code uses the torch lightning framework. Please refer to the documentation of torch lightning (https://lightning.ai/docs/pytorch/stable/) for additional information, such as using different flags when launching the code. +For example, in order to run on GPU 0 please add +--trainer.devices=[0] to the launch script. + +### 4. Citing Archer +``` +@misc{zhou2024archer, + title={ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL}, + author={Yifei Zhou and Andrea Zanette and Jiayi Pan and Sergey Levine and Aviral Kumar}, + year={2024}, + eprint={2402.19446}, + archivePrefix={arXiv}, + primaryClass={cs.LG} +} + diff --git a/SimulateOnEnv.py b/SimulateOnEnv.py new file mode 100644 index 0000000000000000000000000000000000000000..d9bd99e80424e4d85624d4c23bcf42eb6791ddb9 --- /dev/null +++ b/SimulateOnEnv.py @@ -0,0 +1,40 @@ +import torch + +def batch_simulate_on_environment(policy, env, verbose = True): + if verbose: + print("*** In batch_simulate_on_environment ***") + + from Dataset import Trajectory, TrajectoryDataset + from math import ceil + + dataset = TrajectoryDataset() + + trajectories = [Trajectory() for _ in range(env.bsize)] + batch_obs = env.reset() + batch_done = [False,]*env.bsize + while not all(batch_done): + with torch.no_grad(): + actions = policy(batch_obs) + batch_feedback = env.step(actions) + for i, feedback in zip(range(env.bsize), batch_feedback): + if feedback is None: + continue + + next_obs, r, done = feedback + + trajectories[i].append({"observation": batch_obs[i], + "action": actions[i], + "reward": r, + "next_observation": next_obs, + "done": done, + }) + batch_obs[i] = next_obs + batch_done[i] = done + for trajectory in trajectories: + dataset.append_trajectory(trajectory) + print(trajectory.transitions[-1].next_observation) + + dataset.check_consistency() + if verbose: + print("Data Coollection is Complete. Returns: \n", dataset.get_all_trajectory_returns(), "\n with mean: ",dataset.mean_trajectory_return(), "\n" ) + return dataset diff --git a/Tasks.py b/Tasks.py new file mode 100644 index 0000000000000000000000000000000000000000..7042186ffa921e687e41449cf86e9ac675f651a5 --- /dev/null +++ b/Tasks.py @@ -0,0 +1,476 @@ +from lightning import LightningDataModule +import torch.utils.data as data +from Dataset import TrajectoryDataset, EmptyDataset +from SimulateOnEnv import batch_simulate_on_environment +import numpy as np +from copy import deepcopy +import sys +import random + + +def rsa_reward(num_feature, min_turns, conv_turn, gamma=2.0): + """ + Nonlinear normalization function, returns u ∈ [0, 1] + - num_feature = min_turns -> u = 1 + - num_feature = conv_turn -> u = 0 + - The closer to min_turns, the slower it approaches 1 + """ + if num_feature == min_turns: + return 1 + # Normalize to [0,1] + u = (conv_turn - num_feature) / (min_turns - num_feature) + # Keep direction (support num_feature < min_turns) + return max(0, min(1, u**gamma)) + + +class Task(LightningDataModule): + def __init__(self, batch_size: int, n_traj_eval: int, **kwargs): + super().__init__(**kwargs) + self.batch_size = batch_size + self.eval_batch_size = self.batch_size + self.n_traj_eval = n_traj_eval + + # Set Defaults + self.shuffle = True + self.drop_last = True # skips last batch to make sure gradient accumulation works as intended + + def setup(self, stage: str): + raise NotImplementedError + + def train_dataloader(self): + return data.DataLoader( + dataset=self.dataset, + batch_size=self.batch_size, + shuffle=self.shuffle, + drop_last=self.drop_last, + num_workers=8, + pin_memory=True, + persistent_workers=True, + ) + + def val_dataloader(self): + return data.DataLoader( + dataset=EmptyDataset(length=self.n_traj_eval), + batch_size=self.eval_batch_size, + pin_memory=True, + ) + + def get_eval_log(self, **kwargs): + pass + + def teardown(self, stage: str): + # Used to clean-up when the run is finished + pass + + +class TwentyQuestions(Task): + def __init__(self, batch_size: int, n_traj_eval: int, word_list=None, **kwargs): + super().__init__(batch_size, n_traj_eval, **kwargs) + + self.word_list = word_list + self.max_horizon = 20 + + def setup(self, stage: str): + self.dataset = self.read_data() + self.dataset.check_consistency() + print( + "\n *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming" + ) + + def read_data(self): + import json + from Dataset import TrajectoryDataset + + f = open("datasets/20q_train.json") + data = json.load(f) + dataset = TrajectoryDataset() + + for game in data: + assert len(game["lines"]) <= 20 + history = "Questions:\n" # assertion is checked with history = '' + for interaction in game["lines"]: + yesAnswer = interaction[-5:] == " Yes." + noAnswer = interaction[-4:] == " No." + assert yesAnswer or noAnswer + observation = history + + done = ( + True if interaction == game["lines"][-1] else False + ) # if the interaction is the last interaction we are done + reward = 0 if done and game["correct"] else -1 + + if yesAnswer: + action = interaction[:-5] + if noAnswer: + action = interaction[:-4] + + history += interaction + "\n" + dataset.append_observation_action_reward(observation, action, reward) + dataset.append_terminal_observation( + history, + trajectory_info={"correct": game["correct"], "word": game["word"]}, + ) + + dataset.check_consistency() + return dataset + + +class RSAGame(Task): + def __init__( + self, + base_model: str, + batch_size: int, + n_traj_eval: int, + word_list=None, + **kwargs, + ): + super().__init__(batch_size, n_traj_eval, **kwargs) + self.base_model = base_model + + self.word_list = word_list + self.max_horizon = 20 + + def setup(self, stage: str): + self.dataset = self.read_data() + self.dataset.check_consistency() + print( + "\n *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming" + ) + + def read_data(self): + import json + from Dataset import TrajectoryDataset + from rsa_game import get_game_outcome, randomly_convert_game_history_to_query + + with open( + f"rsa/{self.base_model}_sampling_all_targets_results.json" + ) as f: + data = json.load(f) + with open( + "rsa/reasoning_dialogs.json" + ) as f: + for key, value in json.load(f).items(): + instance = {} + instance["history"] = value["dialog"] + instance["target"] = value["target_referent"].split(" ") + instance["min_turns"] = len(value["dialog"]) + instance["max_turns"] = len(instance["target"]) * 2 + instance["object_list"] = value["referent_set"] + data.append(instance) + + dataset = TrajectoryDataset() + + for game in random.sample(data, 3200): + is_valid = True + for message in game["history"]: + if message["content"] == "": + is_valid = False + break + if not is_valid: + continue + + outcome, history_length = get_game_outcome( + game["history"], game["target"], game["min_turns"] + ) + if outcome == "game wins": + reward = rsa_reward( + len(game["target"]) * 2, game["min_turns"] * 2, history_length + ) + else: + continue + + if reward == 0: + continue + + for idx, interaction in enumerate(game["history"][:history_length]): + query = randomly_convert_game_history_to_query( + game["history"][:idx], + target=game["target"], + min_turns=game["min_turns"], + object_list=game["object_list"], + ) + target = interaction["content"] + + done = ( + True if idx >= history_length - 2 else False + ) # if the interaction is the last interaction we are done + reward = 0 if done else reward + + dataset.append_observation_action_reward(query, target, reward) + + history = randomly_convert_game_history_to_query( + game["history"], + target=game["target"], + min_turns=game["min_turns"], + object_list=game["object_list"], + ) + dataset.append_terminal_observation( + history, + trajectory_info={ + "object_list": game["object_list"], + "target": game["target"], + }, + ) + + print("The length of the dataset is: ", len(dataset)) + dataset.check_consistency() + return dataset + + +class WordTaboo(Task): + def __init__( + self, + base_model: str, + batch_size: int, + n_traj_eval: int, + word_list=None, + **kwargs, + ): + super().__init__(batch_size, n_traj_eval, **kwargs) + + self.base_model = base_model + self.word_list = word_list + self.max_horizon = 20 + + def setup(self, stage: str): + self.dataset = self.read_data() + self.dataset.check_consistency() + print( + "\n *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming" + ) + + def read_data(self): + import json + from Dataset import TrajectoryDataset + from word_taboo import get_game_outcome, randomly_convert_game_history_to_query + + with open( + f"wordtaboo/{self.base_model}_sampling_all_targets_results.json", "r" + ) as f: + data = json.load(f) + with open( + "wordtaboo/llm_game_top_test_results.json", "r" + ) as f: + data.extend(json.load(f)) + + dataset = TrajectoryDataset() + + for game in data: + is_valid = True + for message in game["history"]: + if message["content"] == "": + is_valid = False + break + if not is_valid: + continue + + outcome, history_length = get_game_outcome( + game["history"], game["target"], game["max_turns"] + ) + + if outcome == "defender wins": + winner = "defender" + + elif outcome == "attacker wins": + if self.base_model == "Qwen3-14B": + if random.random() < 0.85: # 0.85 for qwen3; 0.9 for llama3 + continue + else: + if random.random() < 0.9: # 0.85 for qwen3; 0.9 for llama3 + continue + winner = "attacker" + + else: + continue + + for idx, interaction in enumerate(game["history"][:history_length]): + if interaction["role"] != winner: + continue + + query = randomly_convert_game_history_to_query( + game["history"][:idx], + target=game["target"], + max_turns=game["max_turns"], + ) + + target = interaction["content"] + + done = ( + True if idx >= history_length - 2 else False + ) # if the interaction is the last interaction we are done + reward = 0 if done else 1 + + dataset.append_observation_action_reward(query, target, reward) + + history = randomly_convert_game_history_to_query( + game["history"], game["target"], game["max_turns"] + ) + dataset.append_terminal_observation( + history, trajectory_info={"target": game["target"]} + ) + print("The length of the dataset is: ", len(dataset)) + dataset.check_consistency() + return dataset + + +class StrategicDialogue(Task): + def __init__( + self, + base_model: str, + batch_size: int, + n_traj_eval: int, + word_list=None, + **kwargs, + ): + super().__init__(batch_size, n_traj_eval, **kwargs) + self.base_model = base_model + + self.word_list = word_list + self.max_horizon = 20 + + def setup(self, stage: str): + self.dataset = self.read_data() + self.dataset.check_consistency() + print( + "\n *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming" + ) + + def read_data(self): + import json + from Dataset import TrajectoryDataset + from word_taboo import get_game_outcome, randomly_convert_game_history_to_query + + with open( + f"wordtaboo/{self.base_model}_sampling_all_targets_results.json", "r" + ) as f: + data = json.load(f) + with open( + "wordtaboo/llm_game_top_test_results.json", "r" + ) as f: + data.extend(json.load(f)) + + dataset = TrajectoryDataset() + + for game in data: + is_valid = True + for message in game["history"]: + if message["content"] == "": + is_valid = False + break + if not is_valid: + continue + + outcome, history_length = get_game_outcome( + game["history"], game["target"], game["max_turns"] + ) + + if outcome == "defender wins": + winner = "defender" + + elif outcome == "attacker wins": + if self.base_model == "Qwen3-14B": + if random.random() < 0.85: # 0.85 for qwen3; 0.9 for llama3 + continue + else: + if random.random() < 0.9: # 0.85 for qwen3; 0.9 for llama3 + continue + winner = "attacker" + + else: + continue + + for idx, interaction in enumerate(game["history"][:history_length]): + if interaction["role"] != winner: + continue + + query = randomly_convert_game_history_to_query( + game["history"][:idx], + target=game["target"], + max_turns=game["max_turns"], + ) + + target = interaction["content"] + + done = ( + True if idx >= history_length - 2 else False + ) # if the interaction is the last interaction we are done + reward = 0 if done else 1 + + dataset.append_observation_action_reward(query, target, reward) + + history = randomly_convert_game_history_to_query( + game["history"], game["target"], game["max_turns"] + ) + dataset.append_terminal_observation( + history, trajectory_info={"target": game["target"]} + ) + + from rsa_game import get_game_outcome, randomly_convert_game_history_to_query + with open( + f"rsa/{self.base_model}_sampling_all_targets_results.json" + ) as f: + data = json.load(f) + with open( + "rsa/reasoning_dialogs.json" + ) as f: + for key, value in json.load(f).items(): + instance = {} + instance["history"] = value["dialog"] + instance["target"] = value["target_referent"].split(" ") + instance["min_turns"] = len(value["dialog"]) + instance["max_turns"] = len(instance["target"]) * 2 + instance["object_list"] = value["referent_set"] + data.append(instance) + + for game in random.sample(data, 3200): + is_valid = True + for message in game["history"]: + if message["content"] == "": + is_valid = False + break + if not is_valid: + continue + + outcome, history_length = get_game_outcome( + game["history"], game["target"], game["min_turns"] + ) + if outcome == "game wins": + reward = rsa_reward( + len(game["target"]) * 2, game["min_turns"] * 2, history_length + ) + else: + continue + + for idx, interaction in enumerate(game["history"][:history_length]): + query = randomly_convert_game_history_to_query( + game["history"][:idx], + target=game["target"], + min_turns=game["min_turns"], + object_list=game["object_list"], + ) + target = interaction["content"] + + done = ( + True if idx >= history_length - 2 else False + ) # if the interaction is the last interaction we are done + reward = 0 if done else reward + + dataset.append_observation_action_reward(query, target, reward) + + history = randomly_convert_game_history_to_query( + game["history"], + target=game["target"], + min_turns=game["min_turns"], + object_list=game["object_list"], + ) + dataset.append_terminal_observation( + history, + trajectory_info={ + "object_list": game["object_list"], + "target": game["target"], + }, + ) + + print("The length of the dataset is: ", len(dataset)) + dataset.check_consistency() + return dataset diff --git a/__pycache__/Algorithms.cpython-310.pyc b/__pycache__/Algorithms.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..af238d3bce78e4c945925914ec0b97c9985dc865 Binary files /dev/null and b/__pycache__/Algorithms.cpython-310.pyc differ diff --git a/__pycache__/ArcherCritic.cpython-310.pyc b/__pycache__/ArcherCritic.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..fe1df8e965c509fd32f91f6159577c741ff0161f Binary files /dev/null and b/__pycache__/ArcherCritic.cpython-310.pyc differ diff --git a/__pycache__/Dataset.cpython-310.pyc b/__pycache__/Dataset.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..664d8d8cc8607005f449e0b2651df7fe8366e2c6 Binary files /dev/null and b/__pycache__/Dataset.cpython-310.pyc differ diff --git a/__pycache__/SimulateOnEnv.cpython-310.pyc b/__pycache__/SimulateOnEnv.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..17efecef035763742931de07131e5006fffc49c0 Binary files /dev/null and b/__pycache__/SimulateOnEnv.cpython-310.pyc differ diff --git a/__pycache__/Tasks.cpython-310.pyc b/__pycache__/Tasks.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..6ca155cdf71bbd8457bb1096396274cc60932bd7 Binary files /dev/null and b/__pycache__/Tasks.cpython-310.pyc differ diff --git a/__pycache__/rsa_game.cpython-310.pyc b/__pycache__/rsa_game.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..cd4d439dd1ff2d0fc97062bcb4361640f891c673 Binary files /dev/null and b/__pycache__/rsa_game.cpython-310.pyc differ diff --git a/__pycache__/twenty_questions.cpython-310.pyc b/__pycache__/twenty_questions.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..f777867842b57629290d22cdbd64ab2dadfdd4fd Binary files /dev/null and b/__pycache__/twenty_questions.cpython-310.pyc differ diff --git a/__pycache__/word_taboo.cpython-310.pyc b/__pycache__/word_taboo.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..84990d10900d2025e8c90e14c37171a2e01d7f3c Binary files /dev/null and b/__pycache__/word_taboo.cpython-310.pyc differ diff --git a/archer.egg-info/PKG-INFO b/archer.egg-info/PKG-INFO new file mode 100644 index 0000000000000000000000000000000000000000..4c9b01d16b15fc006fcc8d94bd4aaa732e1f71cc --- /dev/null +++ b/archer.egg-info/PKG-INFO @@ -0,0 +1,95 @@ +Metadata-Version: 2.4 +Name: archer +Version: 0.1.0 +Summary: Research code for Offline ArCHer (Actor Critic Framework with Hierarchical Structures) +Home-page: https://github.com/andreazanette/OfflineArcher.git +Author: Andrea Zanette +License: MIT +Keywords: ArCHer +Classifier: Intended Audience :: Science/Research +Classifier: License :: OSI Approved :: MIT License +Classifier: Programming Language :: Python :: 3 +Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence +Requires-Python: >=3.7 +Description-Content-Type: text/markdown +Requires-Dist: lightning +Requires-Dist: jsonargparse +Requires-Dist: lightning-utilities +Requires-Dist: numpy +Requires-Dist: pytorch-lightning +Requires-Dist: safetensors +Requires-Dist: sentencepiece +Requires-Dist: torch +Requires-Dist: torchmetrics +Requires-Dist: tqdm +Requires-Dist: transformers +Requires-Dist: wandb +Requires-Dist: jsonargparse[signatures]>=4.26.1 +Dynamic: author +Dynamic: classifier +Dynamic: description +Dynamic: description-content-type +Dynamic: home-page +Dynamic: keywords +Dynamic: license +Dynamic: requires-dist +Dynamic: requires-python +Dynamic: summary + +# OfflineArcher +Research Code for the Offline Experiments of "ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL" + +[Yifei Zhou](https://yifeizhou02.github.io/), [Andrea Zanette](https://azanette.com/), [Jiayi Pan](https://www.jiayipan.me/), [Aviral Kumar](https://aviralkumar2907.github.io/), [Sergey Levine](https://people.eecs.berkeley.edu/~svlevine/) + +![archer_diagram 001](https://github.com/YifeiZhou02/ArCHer/assets/83000332/b874432a-d330-49a5-906c-bba37e17f831) + + +This repo supports the following methods: + +- [Offline ArCHer][1] +- Offline Filtered BC +- Offline BC + +[1]: https://github.com/YifeiZhou02/ArCHer + +And the following environments +- [Twenty Questions][2] + +[2]: https://lmrl-gym.github.io/ + + +## Quick Start +### 1. Install Dependencies +```bash +conda create -n archer python==3.10 +conda activate archer + +git clone https://github.com/andreazanette/OfflineArcher +cd OfflineArcher +python -m pip install -e . +``` +### 2. Download Datasets and Oracles +Offline datasets and Oracles checkpoints used in the paper can be found [here](https://drive.google.com/drive/folders/1pRocQI0Jv479G4vNMtQn1JOq8Shf2B6U?usp=sharing). +You will need to create an "oracles" and "datasets" folder and put the oracle and dataset in such folders. +The oracle for Twenty Questions should be named 20q_t5_oracle.pt and the dataset should be called "twenty_questions.json". + +### 3. Run Experiments +You can directly run experiments by runnig the launch scripts. For example, in order to lauch Offline Archer on Twenty Question simply run +```bash +. submit_OfflineArcher_TwentyQuestions.sh +``` +The code uses the torch lightning framework. Please refer to the documentation of torch lightning (https://lightning.ai/docs/pytorch/stable/) for additional information, such as using different flags when launching the code. +For example, in order to run on GPU 0 please add +--trainer.devices=[0] to the launch script. + +### 4. Citing Archer +``` +@misc{zhou2024archer, + title={ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL}, + author={Yifei Zhou and Andrea Zanette and Jiayi Pan and Sergey Levine and Aviral Kumar}, + year={2024}, + eprint={2402.19446}, + archivePrefix={arXiv}, + primaryClass={cs.LG} +} + diff --git a/archer.egg-info/SOURCES.txt b/archer.egg-info/SOURCES.txt new file mode 100644 index 0000000000000000000000000000000000000000..eb0f8184ea1e99c88d3a93cbc4a885c55bd63cce --- /dev/null +++ b/archer.egg-info/SOURCES.txt @@ -0,0 +1,7 @@ +README.md +setup.py +archer.egg-info/PKG-INFO +archer.egg-info/SOURCES.txt +archer.egg-info/dependency_links.txt +archer.egg-info/requires.txt +archer.egg-info/top_level.txt \ No newline at end of file diff --git a/archer.egg-info/dependency_links.txt b/archer.egg-info/dependency_links.txt new file mode 100644 index 0000000000000000000000000000000000000000..8b137891791fe96927ad78e64b0aad7bded08bdc --- /dev/null +++ b/archer.egg-info/dependency_links.txt @@ -0,0 +1 @@ + diff --git a/archer.egg-info/requires.txt b/archer.egg-info/requires.txt new file mode 100644 index 0000000000000000000000000000000000000000..6bbc7e11afd993cbadcab1a122d76ecb8543d50e --- /dev/null +++ b/archer.egg-info/requires.txt @@ -0,0 +1,13 @@ +lightning +jsonargparse +lightning-utilities +numpy +pytorch-lightning +safetensors +sentencepiece +torch +torchmetrics +tqdm +transformers +wandb +jsonargparse[signatures]>=4.26.1 diff --git a/archer.egg-info/top_level.txt b/archer.egg-info/top_level.txt new file mode 100644 index 0000000000000000000000000000000000000000..8b137891791fe96927ad78e64b0aad7bded08bdc --- /dev/null +++ b/archer.egg-info/top_level.txt @@ -0,0 +1 @@ + diff --git a/archer.tar.gz b/archer.tar.gz new file mode 100644 index 0000000000000000000000000000000000000000..316da98ebb848ddf3747401a23c5e549167f9c33 --- /dev/null +++ b/archer.tar.gz @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56ce34aac81f6fa571d4d52496dff623e49463700b9f3506e0c12a467d56f16e +size 4223243374 diff --git a/checkpoints/archer_Llama3-8B-I_strategic/README.md b/checkpoints/archer_Llama3-8B-I_strategic/README.md new file mode 100644 index 0000000000000000000000000000000000000000..319b561f257156d007acf5bb0c26546d87668321 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/README.md @@ -0,0 +1,207 @@ +--- +base_model: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model +library_name: peft +pipeline_tag: text-generation +tags: +- base_model:adapter:/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model +- lora +- transformers +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.17.1 \ No newline at end of file diff --git a/checkpoints/archer_Llama3-8B-I_strategic/adapter_config.json b/checkpoints/archer_Llama3-8B-I_strategic/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2a54e98297f70779b7b060df97696aa3a5f4b218 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/adapter_config.json @@ -0,0 +1,37 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "qalora_group_size": 16, + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "q_proj", + "v_proj" + ], + "target_parameters": null, + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_qalora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoints/archer_Llama3-8B-I_strategic/adapter_model.safetensors b/checkpoints/archer_Llama3-8B-I_strategic/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29723ae6249d08faa4082315fd2f5c3254e272f4 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb3d93de581e4e6b207ed65fa6970bb0ef8ca2a52e8bd9edebc8ab70945f5039 +size 6832728 diff --git a/checkpoints/archer_Llama3-8B-I_strategic/chat_template.jinja b/checkpoints/archer_Llama3-8B-I_strategic/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..39bd0c9f7fe30aea14eda194fee17703da4a4dbf --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/chat_template.jinja @@ -0,0 +1,5 @@ +{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|> + +'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|> + +' }}{% endif %} \ No newline at end of file diff --git a/checkpoints/archer_Llama3-8B-I_strategic/special_tokens_map.json b/checkpoints/archer_Llama3-8B-I_strategic/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..344c8261025248cbe380e52f8730a03149d599e1 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/special_tokens_map.json @@ -0,0 +1,23 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/checkpoints/archer_Llama3-8B-I_strategic/tokenizer.json b/checkpoints/archer_Llama3-8B-I_strategic/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..648536220968af6dd775017e76bdfe29bd7ccb61 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fc5ed64d17c57f61c0ef996ac8b3a8918e7d406866cc4a0292d362a31a217e4 +size 17210125 diff --git a/checkpoints/archer_Llama3-8B-I_strategic/tokenizer_config.json b/checkpoints/archer_Llama3-8B-I_strategic/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..983965e9a185b9a348465899c8743e3deeaf23bf --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_strategic/tokenizer_config.json @@ -0,0 +1,2065 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_248|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_249|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_250|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 1024, + "pad_token": "<|eot_id|>", + "padding_side": "left", + "tokenizer_class": "PreTrainedTokenizerFast", + "truncation_side": "left" +} diff --git a/checkpoints/archer_Llama3-8B-I_word/README.md b/checkpoints/archer_Llama3-8B-I_word/README.md new file mode 100644 index 0000000000000000000000000000000000000000..8f1d7244e8eeff6139f7f39cbe6392393cc27a40 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/README.md @@ -0,0 +1,207 @@ +--- +base_model: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model +library_name: peft +pipeline_tag: text-generation +tags: +- base_model:adapter:/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model +- lora +- transformers +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.17.1 \ No newline at end of file diff --git a/checkpoints/archer_Llama3-8B-I_word/adapter_config.json b/checkpoints/archer_Llama3-8B-I_word/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7ebd87a526fb0f0813936e360f2f5350c64a35fd --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/adapter_config.json @@ -0,0 +1,37 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 16, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "qalora_group_size": 16, + "r": 8, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "v_proj", + "q_proj" + ], + "target_parameters": null, + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_qalora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/checkpoints/archer_Llama3-8B-I_word/adapter_model.safetensors b/checkpoints/archer_Llama3-8B-I_word/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b0965a3f0006838545ded70cfbaa19243012aea8 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4718a43b261c26cfc0d18d7ce3f147b5010ceeee38d74f1aa3c1f163c9203e6 +size 6832728 diff --git a/checkpoints/archer_Llama3-8B-I_word/chat_template.jinja b/checkpoints/archer_Llama3-8B-I_word/chat_template.jinja new file mode 100644 index 0000000000000000000000000000000000000000..39bd0c9f7fe30aea14eda194fee17703da4a4dbf --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/chat_template.jinja @@ -0,0 +1,5 @@ +{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|> + +'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|> + +' }}{% endif %} \ No newline at end of file diff --git a/checkpoints/archer_Llama3-8B-I_word/special_tokens_map.json b/checkpoints/archer_Llama3-8B-I_word/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..344c8261025248cbe380e52f8730a03149d599e1 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/special_tokens_map.json @@ -0,0 +1,23 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/checkpoints/archer_Llama3-8B-I_word/tokenizer.json b/checkpoints/archer_Llama3-8B-I_word/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..648536220968af6dd775017e76bdfe29bd7ccb61 --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fc5ed64d17c57f61c0ef996ac8b3a8918e7d406866cc4a0292d362a31a217e4 +size 17210125 diff --git a/checkpoints/archer_Llama3-8B-I_word/tokenizer_config.json b/checkpoints/archer_Llama3-8B-I_word/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..983965e9a185b9a348465899c8743e3deeaf23bf --- /dev/null +++ b/checkpoints/archer_Llama3-8B-I_word/tokenizer_config.json @@ -0,0 +1,2065 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_248|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_249|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_250|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 1024, + "pad_token": "<|eot_id|>", + "padding_side": "left", + "tokenizer_class": "PreTrainedTokenizerFast", + "truncation_side": "left" +} diff --git a/deepspeed_zero3.yaml b/deepspeed_zero3.yaml new file mode 100644 index 0000000000000000000000000000000000000000..768bc487357f7fb71bdba7072bcc6d70f47448ac --- /dev/null +++ b/deepspeed_zero3.yaml @@ -0,0 +1,66 @@ +compute_environment: LOCAL_MACHINE +debug: false +deepspeed_config: + deepspeed_multinode_launcher: standard + zero3_save_16bit_model: true + zero_optimization: + stage: 3 + offload_optimizer: + device: cpu + pin_memory: true + buffer_count: 4 + offload_param: + device: cpu + pin_memory: true + buffer_count: 4 + model_parallel: + enable: true + tensor_parallel_size: 4 + pipeline_parallel_size: 1 + reduce_bucket_size: 2e5 + stage3_prefetch_bucket_size: 0.1e6 # Reduced further + stage3_max_live_parameters: 1e6 + stage3_max_reuse_distance: 1e5 + stage3_max_live_gradients: 1e6 + stage3_param_persistence_threshold: 1e4 + sub_group_size: 1e9 + + activation_checkpointing: + partition_activations: true + contiguous_memory_optimization: true + cpu_checkpointing: true + + bf16: + enabled: false + fp16: + enabled: true + + optimizer: + type: AdamW + params: + eight_bit: true + lr: auto + betas: auto + eps: auto + weight_decay: auto + + scheduler: + type: WarmupDecayLR + params: + num_warmup_steps: auto + num_training_steps: auto + + communication_data_type: fp16 + communication_bucket_size: 500000000 + dist_backend: nccl + train_micro_batch_size_per_gpu: auto + gradient_clipping: 1.0 + +distributed_type: DEEPSPEED +downcast_bf16: auto +machine_rank: 0 +main_training_function: main +mixed_precision: bf16 +rdzv_backend: static +same_network: true +use_cpu: auto \ No newline at end of file diff --git a/main.py b/main.py new file mode 100644 index 0000000000000000000000000000000000000000..2f85bddf1fb0fb30c705a98eaa0b6048f51f6203 --- /dev/null +++ b/main.py @@ -0,0 +1,33 @@ +# main.py +from lightning.pytorch.cli import LightningCLI +from lightning.pytorch.callbacks import ModelCheckpoint +from lightning.fabric import seed_everything +from Algorithms import ActorCritic +from Tasks import TwentyQuestions +import torch +torch.backends.cuda.matmul.allow_tf32 = True # 启用 TF32 加速 +torch.backends.cudnn.allow_tf32 = True +from torch.cuda.amp import GradScaler # 如果用 AMP + +seed_everything(42) + +def cli_main(): + checkpoint_callback = ModelCheckpoint( + dirpath="models", + every_n_train_steps=200, + every_n_epochs=0, + save_top_k=1, + save_last=True # 同时保存最后一个模型 + ) + cli = LightningCLI( + save_config_kwargs={"overwrite": True}, + trainer_defaults={ # 添加 trainer 默认配置 + "callbacks": [checkpoint_callback], + "limit_val_batches": 0, + "val_check_interval": None + } + ) + exit(0) + +if __name__ == "__main__": + cli_main() \ No newline at end of file diff --git a/requirements.txt b/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..6bbc7e11afd993cbadcab1a122d76ecb8543d50e --- /dev/null +++ b/requirements.txt @@ -0,0 +1,13 @@ +lightning +jsonargparse +lightning-utilities +numpy +pytorch-lightning +safetensors +sentencepiece +torch +torchmetrics +tqdm +transformers +wandb +jsonargparse[signatures]>=4.26.1 diff --git a/rsa_game.py b/rsa_game.py new file mode 100644 index 0000000000000000000000000000000000000000..aa526563957be8e0a110b96e27296fb8d97d3e3d --- /dev/null +++ b/rsa_game.py @@ -0,0 +1,383 @@ +import re +import string +import random +from typing import Optional, Tuple, List, Dict +from nltk.stem import PorterStemmer +PSTEMMER = PorterStemmer() + +GAME_RULE_PROMPTS = [ + """Play the game of Collabrative Rational Speech Act Game. In this game, there are two players, a speaker and a listener. + +At the beginning, the speaker is assigned a target object, with which the listener is not informed. The task of the speaker is to guide the listener to guess the target object, then they win the game. However, the speaker is only allowed to give one feature per turn. + +At the same time, the listener tries to figure out the target object and give the possible target referent objects at each turn. The listener can give more than one possible target referent object sets at each turn. If the listener identifies the target object, he can say "I know the target object! It is `target object`!". + +At each turn, the speaker should try to give a feature of the target object which provides the listener with the most information, and the listener would update the possible target referent objects from the previous turn. + +Remember, the listener can only update his referent set from the previous turn's guess, he cannot add new referents. + +The less turns they take to guess the target object, the higher the score they get. +""", + """Engage in the collaborative challenge of Rational Speech Act Game, featuring two participants: one takes on the role of the speaker, while the other serves as the listener. + +Initially, the speaker is secretly assigned a target object, which remains unknown to the listener. The speaker's objective is to strategically guide the listener toward identifying the target object, thereby securing victory. However, there's a constraint: the speaker may only provide one feature per turn. + +Simultaneously, the listener's mission is to deduce the target object and present possible target referent objects at each turn. The listener has the flexibility to offer multiple possible target referent object sets during their turn. If the listener identifies the target object, they can declare "I know the target object! It is `target object`!". + +During each turn, the speaker should aim to provide a feature of the target object that maximizes the information available to the listener, while the listener updates their possible target referent objects based on the previous turn's information. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring system rewards efficiency: fewer turns required to guess the target object results in higher scores. +""", + """Dive into the strategic collaboration known as Rational Speech Act Game, where two players assume distinct roles: the speaker and the listener. + +To commence the game, the speaker receives a target object in secret, while the listener remains unaware of this object. The speaker's challenge is to effectively guide the listener toward correctly identifying the target object, which would result in a win. However, there's a limitation: the speaker can only share one feature per turn. + +Concurrently, the listener embarks on a quest to determine the target object and must present possible target referent objects at each turn. The listener enjoys the advantage of being able to propose multiple possible target referent object sets during their turn. If the listener identifies the target object, they can proclaim "I know the target object! It is `target object`!". + +At each turn, the speaker should strategically select a feature of the target object that provides the listener with maximum informational value, while the listener refines their possible target referent objects based on the previous turn's revelations. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring mechanism emphasizes efficiency: achieving the target object identification in fewer turns yields higher scores. +""", + """Step into the cooperative challenge of Rational Speech Act Game, a game designed for two players: one acting as the speaker, the other as the listener. + +In the opening phase, the speaker is discreetly given a target object, which is kept hidden from the listener. The speaker's goal is to successfully lead the listener to identify the target object, thereby claiming victory. However, there's a rule: the speaker is restricted to providing only one feature per turn. + +At the same time, the listener's task is to figure out the target object and present possible target referent objects at each turn. The listener has the liberty to suggest multiple possible target referent object sets during their turn. If the listener identifies the target object, they can announce "I know the target object! It is `target object`!". + +During each turn, the speaker should carefully choose a feature of the target object that delivers the most valuable information to the listener, while the listener adjusts their possible target referent objects based on the previous turn's insights. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring system prioritizes efficiency: the fewer turns taken to identify the target object, the higher the score achieved. +""", + """Immerse yourself in the collaborative strategy game called Rational Speech Act Game, featuring two distinct roles: the speaker and the listener. + +As the game begins, the speaker is covertly assigned a target object, which remains a mystery to the listener. The speaker's mission is to skillfully guide the listener toward correctly identifying the target object, which would secure a win. However, there's a restriction: the speaker may only offer one feature per turn. + +Meanwhile, the listener is engaged in a process of deduction, attempting to determine the target object and presenting possible target referent objects at each turn. The listener benefits from the ability to suggest multiple possible target referent object sets during their turn. If the listener identifies the target object, they can state "I know the target object! It is `target object`!". + +At each turn, the speaker should strategically provide a feature of the target object that maximizes the informational benefit for the listener, while the listener updates their possible target referent objects based on the previous turn's information. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring framework rewards efficiency: fewer turns required to identify the target object results in higher scores. +""", + """Dive into the collaborative challenge known as Rational Speech Act Game, where two players take on specific roles: the speaker and the listener. + +The speaker starts the game with a secret target object assignment, while the listener remains in the dark about this object. The speaker's objective is to effectively guide the listener toward identifying the target object, thereby achieving victory. However, there's a constraint: the speaker can only provide one feature per turn. + +Concurrently, the listener's challenge is to deduce the target object and present possible target referent objects at each turn. The listener enjoys the flexibility of being able to suggest multiple possible target referent object sets during their turn. If the listener identifies the target object, they can express "I know the target object! It is `target object`!". + +During each turn, the speaker should aim to provide a feature of the target object that offers the listener the most valuable information, while the listener refines their possible target referent objects based on the previous turn's revelations. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring mechanism emphasizes efficiency: achieving target object identification in fewer turns yields higher scores. +""", + """Step into the strategic collaboration called Rational Speech Act Game, where the roles of speaker and listener are central to the gameplay. + +The speaker embarks on this journey with a secretly assigned target object, while the listener begins unaware of this object. The speaker's challenge is to successfully guide the listener toward identifying the target object, which would result in a win. However, there's a limitation: the speaker is only permitted to share one feature per turn. + +On the other side, the listener's mission is to figure out the target object and present possible target referent objects at each turn. The listener has the advantage of being able to propose multiple possible target referent object sets during their turn. If the listener identifies the target object, they can reveal "I know the target object! It is `target object`!". + +At each turn, the speaker should strategically select a feature of the target object that provides the listener with maximum informational value, while the listener updates their possible target referent objects based on the previous turn's insights. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring system prioritizes efficiency: fewer turns taken to identify the target object results in higher scores. +""", + """Embark on the collaborative challenge of Rational Speech Act Game, where players assume the roles of either speaker or listener. + +The speaker enters the game with a covertly assigned target object, while the listener starts without knowledge of this object. The speaker's goal is to effectively guide the listener toward identifying the target object, thereby securing victory. However, there's a rule: the speaker may only provide one feature per turn. + +Simultaneously, the listener's task is to deduce the target object and present possible target referent objects at each turn. The listener benefits from the ability to suggest multiple possible target referent object sets during their turn. If the listener identifies the target object, they can declare "I know the target object! It is `target object`!". + +During each turn, the speaker should carefully choose a feature of the target object that delivers the most valuable information to the listener, while the listener adjusts their possible target referent objects based on the previous turn's information. + +Remember, the listener can only update their referent set from the previous turn's guess; they cannot add new referents. + +The scoring framework rewards efficiency: the fewer turns required to identify the target object, the higher the score achieved. +""", +] + + +INSTRUCT_PROMPTS = { + "speaker": """\n\n### Instruction: You are the pragmatic rational speaker. The target object is `{target}` and the object list is '{object_list}'. Provide your response including the object feature.\n\n### Response:""", + "listener": """\n\n### Instruction: Your are the pragmatic rational listener. The object list is '{object_list}'. Provide your infered target object or the possible target object sets.\n\n### Response:""", +} + +PLAYER_INSTRUCT_PROMPTS = { + "speaker": "You are the pragmatic rational speaker. The target object is `{target}` and the object list is '{object_list}'. Provide your response including the object feature.", + "listener": "Your are the pragmatic rational listener. The object list is '{object_list}'. Provide your infered target object or the possible target object sets.", +} + +ROLES = ["speaker", "listener"] +PREDICT_TEMP = r"i know the target object" + +def calculate_reward(turn, golden_turn): + if turn <= golden_turn: + return 1 + else: + return 0 + + +def get_object_list(formatted_str): + """Reverse the item_string function to extract the original list of items""" + # If the string is empty, return empty list + if not formatted_str: + return [] + + # Remove surrounding brackets/parentheses/braces if present + if formatted_str.startswith("[") and formatted_str.endswith("]"): + formatted_str = formatted_str[1:-1].strip() + elif formatted_str.startswith("(") and formatted_str.endswith(")"): + formatted_str = formatted_str[1:-1].strip() + elif formatted_str.startswith("{") and formatted_str.endswith("}"): + formatted_str = formatted_str[1:-1].strip() + + # Handle different formats + if ", and " in formatted_str: # Oxford comma format + parts = formatted_str.split(", and ") + if len(parts) > 1: + return [item.strip() for item in parts[0].split(", ")] + [parts[1].strip()] + elif " and " in formatted_str and ", " in formatted_str: # No Oxford comma + parts = formatted_str.split(", ") + last_part = parts[-1] + if " and " in last_part: + last_items = last_part.split(" and ") + return [item.strip() for item in parts[:-1]] + [ + item.strip() for item in last_items + ] + elif " and " in formatted_str: # "and" separated format + return [item.strip() for item in formatted_str.split(" and ")] + elif ", " in formatted_str: # Comma separated format + return [item.strip() for item in formatted_str.split(", ")] + + # If none of the above patterns match, return the string as a single-item list + return [formatted_str.strip()] + + +def get_derivative_targets(object_list, random_choice=True): + """Format a list of items in various natural ways""" + # Randomly choose from different formats + format_choices = { + "comma": lambda x: ", ".join(x), + "bracket": lambda x: "[" + ", ".join(x) + "]", + "parenthesis": lambda x: "(" + ", ".join(x) + ")", + "curly": lambda x: "{" + ", ".join(x) + "}", + "oxford_comma": lambda x: ", ".join(x[:-1]) + ", and " + x[-1], + "and_comma": lambda x: " and ".join(x[:-1]) + ", and " + x[-1], + "no_oxford_comma": lambda x: ", ".join(x[:-1]) + " and " + x[-1], + "and_separated": lambda x: " and ".join(x), + } + + if random_choice: + format = random.choice(list(format_choices.keys())) + return format_choices[format](object_list) + else: + return [func(object_list) for format, func in format_choices.items()] + + +def has_target_object(content: str, target_object: list): + derivative_objects = get_derivative_targets(target_object, random_choice=False) + return any([object in content.lower() for object in derivative_objects]) + + +def is_prediction(content: str, target_object: list): + if re.search(PREDICT_TEMP, content.lower()): + return True + else: + return False + + +def is_correct_prediction(content: str, target_object: list): + derivative_objects = get_derivative_targets(target_object, random_choice=False) + if any(object in content.lower() for object in derivative_objects): + return True + + # Remove all punctuation from content except '.', '!', and ',' + content = ''.join([c for c in content if not (c in set(string.punctuation) and c not in {'.', '!', ','})]) + for feature in target_object: + if feature in content.lower(): + continue + else: + # Use a custom stemmer to handle cases like "crispy" -> "crisp" + def custom_stem(word): + if word.endswith("y") and len(word) > 3: + return PSTEMMER.stem(word[:-1]) + return PSTEMMER.stem(word) + feature_stem = custom_stem(feature.lower()) + content_words = [word for word in content.lower().split()] + content_stems = [custom_stem(word) for word in content_words] + if not any(feature_stem == stem or feature_stem in stem for stem in content_stems): + return False + return True + + # derivative_objects = get_derivative_targets(target_object, random_choice=False) + # predict_regex = [PREDICT_TEMP + object for object in derivative_objects] + # if any([re.search(temp, content.lower()) for temp in predict_regex]): + # return True + # else: + # return False + + +def get_game_outcome(history: List[Dict[str, str]], target_object: list, min_turns: int): + history_length = 0 + for i, item in enumerate(history): + history_length += 1 + if item["role"] == "listener": + if is_prediction(item["content"], target_object): + if is_correct_prediction(item["content"], target_object): + return "game wins", history_length + else: + return "game over", history_length + elif not has_target_object(item["content"], target_object) and i == len(history) - 1: + return "listener does not infer the target object", history_length + else: + if is_prediction(item["content"], target_object): + return "speaker breaks the rules", history_length + + return "game over", history_length + + +def convert_game_history_to_query(history, target, object_list): + GAME_RULE_PROMPT = GAME_RULE_PROMPTS[0] + + history_str = "" + for i, message in enumerate(history): + history_str += "\n - {}: {}".format( + message["role"], message["content"]) + + if len(history) == 0: + query = ( + GAME_RULE_PROMPT + + + "The game is just initialized." + ) + next_player = ROLES[0] + else: + query = ( + GAME_RULE_PROMPT + + "\n### Game History:" + + history_str + ) + if history[-1]["role"] == ROLES[0]: + next_player = ROLES[1] + else: + next_player = ROLES[0] + + query += INSTRUCT_PROMPTS[next_player].format( + target=get_derivative_targets(target), + object_list=get_derivative_targets(object_list), + ) + + return query + + +def randomly_convert_game_history_to_query(history, target, object_list, **kwargs): + if "random_seed" in kwargs: + random.seed(kwargs["random_seed"]) + + if len(history) == 0: + next_player = ROLES[0] + else: + if history[-1]["role"] == ROLES[0]: + next_player = ROLES[1] + else: + next_player = ROLES[0] + + dialog_prefix = "\n" + random.choice( + ["\n - ", "\n### ", "\n## ", "\n# ", "\n *** ", "\n **", "\n\n"] + ) + answer_str, question_str = random.choice( + [ + (next_player, ROLES[1] if next_player == ROLES[0] else ROLES[0]), + ("Assistant", "Human"), + ("Answer", "Question"), + ("Response", "Query"), + ("A", "Q"), + ] + ) + + player_prefix = { + ROLES[0]: answer_str if next_player == ROLES[0] else question_str, + ROLES[1]: answer_str if next_player == ROLES[1] else question_str, + } + + history_str = "" + for i, message in enumerate(history): + history_str += "{}{}: {}".format( + dialog_prefix, player_prefix[message["role"]], message["content"] + ) + + prompt_type = random.choice(["chat", "chat_inverse", "alpaca"]) + system_prefix = random.choice(["Rules", "Game Rule", "System"]) + + system_prompt = random.choice(GAME_RULE_PROMPTS) + + if isinstance(object_list, str): + objects_str = object_list + else: + objects_str = get_derivative_targets(object_list) + if "chat" in prompt_type: + system_prompt += "\n\n" + PLAYER_INSTRUCT_PROMPTS[next_player].format( + target=get_derivative_targets(target), + object_list=objects_str, + ) + + if len(history) == 0: + history_str = "" + system_prompt += "The game is just initialized. " + + system_str = f"{dialog_prefix}{system_prefix}: {system_prompt}" + if "inverse" in prompt_type: + query = ( + history_str + + system_str + + dialog_prefix + + player_prefix[next_player] + + ": " + ) + else: + query = ( + system_str + + history_str + + dialog_prefix + + player_prefix[next_player] + + ": " + ) + + elif prompt_type == "alpaca": + if random.uniform(0, 1) < 0.2: + system_prompt = system_prefix + ": " + system_prompt + + if len(history) == 0: + query = system_prompt + "The game is just initialized. " + else: + query = ( + system_prompt + dialog_prefix + "Game History:" + history_str + "\n\n" + ) + + if random.uniform(0, 1) < 0.2: + query += ( + PLAYER_INSTRUCT_PROMPTS[next_player].format( + target=target, + object_list=objects_str, + )[:-1] + + ": " + ) + else: + query += ( + PLAYER_INSTRUCT_PROMPTS[next_player].format( + target=target, + object_list=objects_str, + ) + + dialog_prefix + + player_prefix[next_player] + + ": " + ) + + return query diff --git a/setup.py b/setup.py new file mode 100644 index 0000000000000000000000000000000000000000..39abf7819fa3ba38a94042ec008447da2be378cd --- /dev/null +++ b/setup.py @@ -0,0 +1,24 @@ +import sys +import setuptools + +setuptools.setup( + name="archer", + version='0.1.0', + url="https://github.com/andreazanette/OfflineArcher.git", + author=("Andrea Zanette"), + description="Research code for Offline ArCHer (Actor Critic Framework with Hierarchical Structures)", + long_description=open("README.md", "r", encoding='utf-8').read(), + long_description_content_type="text/markdown", + keywords='ArCHer', + license='MIT', + packages=setuptools.find_packages(), + install_requires=open("requirements.txt", "r").read().split(), + include_package_data=True, + python_requires='>=3.7', + classifiers=[ + 'Intended Audience :: Science/Research', + 'License :: OSI Approved :: MIT License', + 'Programming Language :: Python :: 3', + 'Topic :: Scientific/Engineering :: Artificial Intelligence', + ], +) \ No newline at end of file diff --git a/submit_BC_TwentyQuestions.sh b/submit_BC_TwentyQuestions.sh new file mode 100644 index 0000000000000000000000000000000000000000..daf6d5abd5b01993212c337259bcdb8b465aa661 --- /dev/null +++ b/submit_BC_TwentyQuestions.sh @@ -0,0 +1,25 @@ +#!/bin/bash -l +# SLURM SUBMIT SCRIPT + +#SBATCH --nodelist=node-gpu01 +#SBATCH --gres=gpu:1 # Request N GPUs per machine + +. init.sh + +lr=1e-4 +batch_size=32 +accumulate_grad_batches=4 #8 + +python main.py fit \ +--data=TwentyQuestions \ +--data.batch_size=$batch_size \ +--data.n_traj_eval=64 \ +--model=BehaviouralCloning \ +--model.lr=$lr \ +--trainer.fast_dev_run=False \ +--trainer.max_epoch=10 \ +--trainer.accumulate_grad_batches=$accumulate_grad_batches \ +--trainer.logger=WandbLogger \ +--trainer.logger.init_args.project="TwentyQuestions-Official" \ +--trainer.logger.init_args.name="BC-lr$lr" \ +--trainer.val_check_interval=500 \ No newline at end of file diff --git a/submit_FBC_TwentyQuestions.sh b/submit_FBC_TwentyQuestions.sh new file mode 100644 index 0000000000000000000000000000000000000000..fbde31e1f5ae43470b8c0768f153db3b8363aa08 --- /dev/null +++ b/submit_FBC_TwentyQuestions.sh @@ -0,0 +1,26 @@ +#!/bin/bash -l +# SLURM SUBMIT SCRIPT + +#SBATCH --nodelist=node-gpu01 +#SBATCH --gres=gpu:1 # Request N GPUs per machine + +. init.sh + +lr=1e-4 +batch_size=32 +accumulate_grad_batches=4 #8 + +python main.py fit \ +--data=TwentyQuestions \ +--data.batch_size=$batch_size \ +--data.n_traj_eval=64 \ +--model=FilteredBehaviouralCloning \ +--model.lr=$lr \ +--model.filter=0.1 \ +--trainer.fast_dev_run=False \ +--trainer.max_epoch=100 \ +--trainer.accumulate_grad_batches=$accumulate_grad_batches \ +--trainer.logger=WandbLogger \ +--trainer.logger.init_args.project="TwentyQuestions-Official" \ +--trainer.logger.init_args.name="FBC-lr$lr" \ +--trainer.val_check_interval=500 \ No newline at end of file diff --git a/submit_OfflineArcher_RSA.sh b/submit_OfflineArcher_RSA.sh new file mode 100644 index 0000000000000000000000000000000000000000..81f56db5b202eff216c445c6f681d69b0f435dc9 --- /dev/null +++ b/submit_OfflineArcher_RSA.sh @@ -0,0 +1,166 @@ +#!/bin/bash -l +# SLURM SUBMIT SCRIPT + +#SBATCH --nodelist=node-gpu01 +#SBATCH --gres=gpu:1 # Request N GPUs per machine + +export TMPDIR=$HOME/tmp +export PYTORCH_CUDA_ALLOC_CONF=max_split_size_mb:64,expandable_segments:True +export NCCL_P2P_DISABLE=1 +export CUDA_LAUNCH_BLOCKING=1 +export CUDA_DEVICE_MAX_CONNECTIONS=1 +export NCCL_NSOCKS_PERTHREAD=4 +export NCCL_SOCKET_NTHREADS=2 +export PYTORCH_NO_CUDA_MEMORY_CACHING=1 # 禁用缓存 +export MAX_JOBS=4 # 限制并行加载 +export ACCELERATE_USE_DEEPSPEED="true" +export TORCH_USE_CUDA_DSA=1 + + +actor_lr=1e-5 +critic_lr=1e-5 + +critic_expectile=0.9 +inv_temp=1.0 + +batch_size=2 +accumulate_grad_batches=16 #8 + +echo "当前conda环境: $(which python)" +echo "python版本: $(python --version)" +echo "accelerate版本: $(accelerate --version)" +echo "CUDA_VISIBLE_DEVICES: $CUDA_VISIBLE_DEVICES" +echo "nvidia-smi:" +nvidia-smi + +echo "==== 测试python能否运行 ====" +python -c "print('Python可以运行')" + +echo "==== 测试main.py是否存在 ====" +ls -l "$HOME/codes/OfflineArcher/main.py" + +echo "==== 测试accelerate能否import torch ====" +accelerate launch --help || echo "accelerate launch命令无法运行" + +echo "==== 开始正式训练 ====" + +set -x # 打印每一行命令,方便debug + +accelerate launch \ + --config_file "$HOME/codes/OfflineArcher/deepspeed_zero3.yaml" \ + --num_processes 4 \ + --gpu_ids 4,5,6,7 \ + --main_process_port 29500 \ + "$HOME/codes/OfflineArcher/main.py" \ + fit \ + --data=RSAGame \ + --data.batch_size=2 \ + --data.base_model="Qwen3-14B" \ + --data.n_traj_eval=4 \ + --model=OfflineArcher \ + --model.optimize_critic=True \ + --model.actor_lr=$actor_lr \ + --model.critic_lr=$critic_lr \ + --model.discount_factor=0.99 \ + --model.tau=0.05 \ + --model.critic_expectile=$critic_expectile \ + --model.inv_temp=$inv_temp \ + --model.accumulate_grad_batches=4 \ + --model.model_name_or_path="/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model" \ + --trainer.fast_dev_run=False \ + --trainer.max_epoch=1 \ + --trainer.logger=WandbLogger \ + --trainer.logger.init_args.project="WordTaboo-Official" \ + --trainer.default_root_dir="checkpoints/archer_Qwen3-14B_word" \ + --trainer.logger.init_args.name="Test-AC-critic_expectile_$critic_expectile-inv_temp_$inv_temp" \ + --trainer.strategy=deepspeed_stage_3 \ + --trainer.devices=4 \ + --trainer.accelerator=gpu \ + --trainer.precision=bf16 \ + --trainer.enable_model_summary=false \ + --trainer.val_check_interval=null \ + --trainer.limit_val_batches=0 > Qwen3-14B_RSA_log.txt 2>&1 + +echo "第三个任务返回码: $?" +tail -20 Qwen3-14B_RSA_log.txt + + +accelerate launch \ + --config_file "$HOME/codes/OfflineArcher/deepspeed_zero3.yaml" \ + --num_processes 4 \ + --gpu_ids 4,5,6,7 \ + --main_process_port 29500 \ + "$HOME/codes/OfflineArcher/main.py" \ + fit \ + --data=WordTaboo \ + --data.batch_size=2 \ + --data.base_model="Qwen3-14B" \ + --data.n_traj_eval=4 \ + --model=OfflineArcher \ + --model.optimize_critic=True \ + --model.actor_lr=$actor_lr \ + --model.critic_lr=$critic_lr \ + --model.discount_factor=0.99 \ + --model.tau=0.05 \ + --model.critic_expectile=$critic_expectile \ + --model.inv_temp=$inv_temp \ + --model.accumulate_grad_batches=4 \ + --model.model_name_or_path="/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model" \ + --trainer.fast_dev_run=False \ + --trainer.max_epoch=1 \ + --trainer.logger=WandbLogger \ + --trainer.logger.init_args.project="WordTaboo-Official" \ + --trainer.default_root_dir="checkpoints/archer_Qwen3-14B_word" \ + --trainer.logger.init_args.name="Test-AC-critic_expectile_$critic_expectile-inv_temp_$inv_temp" \ + --trainer.strategy=deepspeed_stage_3 \ + --trainer.devices=4 \ + --trainer.accelerator=gpu \ + --trainer.precision=bf16 \ + --trainer.enable_model_summary=false \ + --trainer.val_check_interval=null \ + --trainer.limit_val_batches=0 > Qwen3-14B_Word_log.txt 2>&1 + +echo "第三个任务返回码: $?" +tail -20 Qwen3-14B_Word_log.txt + +accelerate launch \ + --config_file "$HOME/codes/OfflineArcher/deepspeed_zero3.syaml" \ + --num_processes 4 \ + --gpu_ids 4,5,6,7 \ + --main_process_port 29500 \ + "$HOME/codes/OfflineArcher/main.py" \ + fit \ + --data=StrategicDialogue \ + --data.batch_size=2 \ + --data.base_model="Qwen3-14B" \ + --data.n_traj_eval=4 \ + --model=OfflineArcher \ + --model.optimize_critic=True \ + --model.actor_lr=$actor_lr \ + --model.critic_lr=$critic_lr \ + --model.discount_factor=0.99 \ + --model.tau=0.05 \ + --model.critic_expectile=$critic_expectile \ + --model.inv_temp=$inv_temp \ + --model.accumulate_grad_batches=4 \ + --model.model_name_or_path="/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model" \ + --trainer.fast_dev_run=False \ + --trainer.max_epoch=1 \ + --trainer.logger=WandbLogger \ + --trainer.logger.init_args.project="Strategic-Official" \ + --trainer.default_root_dir="checkpoints/archer_Qwen3-14B_strategic" \ + --trainer.logger.init_args.name="Test-AC-critic_expectile_$critic_expectile-inv_temp_$inv_temp" \ + --trainer.strategy=deepspeed_stage_3 \ + --trainer.devices=4 \ + --trainer.accelerator=gpu \ + --trainer.precision=bf16 \ + --trainer.enable_model_summary=false \ + --trainer.val_check_interval=null \ + --trainer.limit_val_batches=0 > Qwen3-14B_Strategic_log.txt 2>&1 + +echo "第五个任务返回码: $?" +tail -20 Qwen3-14B_Strategic_log.txt + +set +x + +echo "所有任务执行完毕。请检查上面各个log文件的最后20行和返回码。" \ No newline at end of file diff --git a/submit_OfflineArcher_TwentyQuestions.sh b/submit_OfflineArcher_TwentyQuestions.sh new file mode 100644 index 0000000000000000000000000000000000000000..3bcb39152756b45a1fd1d8ae9e6d5119df6fc794 --- /dev/null +++ b/submit_OfflineArcher_TwentyQuestions.sh @@ -0,0 +1,37 @@ +#!/bin/bash -l +# SLURM SUBMIT SCRIPT + +#SBATCH --nodelist=node-gpu01 +#SBATCH --gres=gpu:1 # Request N GPUs per machine + +. init.sh + +actor_lr=1e-4 +critic_lr=1e-5 + +critic_expectile=0.9 +inv_temp=1.0 + +batch_size=32 +accumulate_grad_batches=4 #8 + +python main.py fit \ +--data=TwentyQuestions \ +--data.batch_size=$batch_size \ +--data.n_traj_eval=64 \ +--model=OfflineArcher \ +--model.optimize_critic=True \ +--model.actor_lr=$actor_lr \ +--model.critic_lr=$critic_lr \ +--model.discount_factor=0.99 \ +--model.tau=0.05 \ +--model.critic_expectile=$critic_expectile \ +--model.inv_temp=$inv_temp \ +--model.accumulate_grad_batches=$accumulate_grad_batches \ +--trainer.fast_dev_run=False \ +--trainer.max_epoch=10 \ +--trainer.logger=WandbLogger \ +--trainer.logger.init_args.project="TwentyQuestions-Official" \ +--trainer.logger.init_args.name="Test-AC-critic_expectile_$critic_expectile-inv_temp_$inv_temp" \ +--trainer.strategy='ddp_find_unused_parameters_true' \ +--trainer.val_check_interval=250 \ No newline at end of file diff --git a/test.py b/test.py new file mode 100644 index 0000000000000000000000000000000000000000..dad778854d0efd2aa7c54cb92ad5fca9fe2c8648 --- /dev/null +++ b/test.py @@ -0,0 +1,34 @@ +from peft import PeftModel, PeftConfig +from transformers import AutoModelForCausalLM, AutoTokenizer +import torch +import os + +# Configuration - replace these with your actual paths +LORA_MODEL_PATH = "./checkpoints/archer_Qwen3-14B_rsa" # e.g., "./lora_output" +BASE_MODEL_NAME = "/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model" # e.g., "./merged_model" + +def merge_lora(base_model_name, lora_path, output_path): + # Load base model + base_model = AutoModelForCausalLM.from_pretrained( + base_model_name, + return_dict=True, + torch_dtype=torch.float16, + device_map="auto" + ) + + # Load tokenizer + tokenizer = AutoTokenizer.from_pretrained(base_model_name) + + # Load LoRA adapter + lora_model = PeftModel.from_pretrained(base_model, lora_path) + + # Merge weights + merged_model = lora_model.merge_and_unload() + + # Save merged model + merged_model.save_pretrained(output_path) + tokenizer.save_pretrained(output_path) + print(f"Merged model saved to {output_path}") + +if __name__ == "__main__": + merge_lora(BASE_MODEL_NAME, LORA_MODEL_PATH, os.path.join(LORA_MODEL_PATH, "merged_model")) \ No newline at end of file diff --git a/twenty_questions.py b/twenty_questions.py new file mode 100644 index 0000000000000000000000000000000000000000..221a3b7049b182212ef3839c92f25ee52b01a9da --- /dev/null +++ b/twenty_questions.py @@ -0,0 +1,149 @@ +import random +from typing import Optional, Dict +import time +import logging +logging.getLogger().setLevel(logging.CRITICAL) +import torch +from transformers import BartTokenizer, BartForConditionalGeneration +from transformers import T5Tokenizer, T5ForConditionalGeneration +import concurrent.futures + +PROMPT_TEMPLATE = 'You are playing a game called twenty questions with me. The rule of twenty question is that you are given a hidden word, and I am guessing what the word is within twenty questions. For every question, if it is an invalid question, you should answer "Invalid Question.". For any valid question, you should answer either "Yes." or "No.". Now the hidden word given to you is "{word}", and the question for the current round is "{question}". Your response is:' +DEFAULT_OBJECT_DICT = { + "Sports": ["Basketball", "Football", "Baseball", "Soccer ball", "Golf ball", "Tennis ball", "Volleyball", "Tennis racket", "Baseball bat", "Helmet"], + "Animals": ["Cat", "Dog", "Horse", "Cow", "Sheep", "Rabbit", "Lion", "Tiger", "Bear", "Elephant"], + "Fruits": ["Apple", "Banana", "Orange", "Strawberry", "Grape", "Watermelon", "Pineapple", "Mango", "Cantaloupe", "Peach"], + "Vehicles": ["Car", "Truck", "Motorcycle", "Boat", "Airplane;Plane", "Train", "Bus", "Helicopter", "Scooter", "Ship"], + "Clothes": ["Shirt", "Pants;Pant;Pair of pants", "Jacket", "Dress", "Skirt", "Belt", "Shoes;Shoe;Pair of shoes", "Boots;Boot;Pair of boots", "Socks;Sock;Pair of socks", "Hat", "Scarf"], + "Electronics": ["Computer", "Smartphone", "Television;TV", "Headphone;Headphones;Pair of headphones", "Monitor;Computer monitor", "Camera", "Microwave;Microwave oven", "Refrigerator", "Blender", "Computer keyboard;Keyboard"], + "Musical Instruments": ["Piano", "Guitar", "Drum;Drums", "Violin", "Saxophone", "Flute", "Trumpet", "Clarinet", "Harp", "Trombone"], + "Furniture": ["Chair", "Table", "Bed", "Desk", "Couch", "Dresser", "Bookcase", "Nightstand", "Mattress", "Pillow"], + "Office Supplies": ["Pen", "Paper;Piece of paper", "Stapler", "Printer", "Calculator", "Battery;Battery pack;Pack of batteries", "Toothbrush", "Toothpaste", "Pencil", "Sharpie", "Scissors;Pair of scissors", "Key", "Diary", "Calendar"], + "Vegetables": ["Carrot", "Potato", "Broccoli", "Tomato", "Onion", "Spinach", "Corn", "Peas;Pea", "Celery", "Cucumber"], + "Art": ["Painting;Canvas painting;Oil painting;Watercolor painting", "Paintbrush", "Canvas;Painting canvas", "Eraser;Pencil eraser", "Marker", "Glue;Glue stick;Bottle of glue", "Sculpture"], + "Kitchen Tools": ["Knife", "Spoon", "Fork", "Plate", "Bowl", "Cooking pot;Pot", "Pan;Saucepan;Frying pan", "Cup", "Chopstick;Chopsticks;Pair of chopsticks", "Whisk"], + "Nature": ["Rock", "Tree", "Bush", "Mountain", "Forest", "Ocean", "Sea", "Lake", "River", "Meteorite", "Cactus"], + "Toys": ["Lego;Lego set", "Doll;Toy doll;Plush doll", "Kite", "Puzzle;Jigsaw puzzle"], + "Jewelry": ["Earring;Earrings;Pair of earrings", "Necklace", "Bracelet", "Ring", "Brooch", "Hairclip", "Pendant", "Watch", "Locket"], + "Garden Supplies": ["Gloves;Glove;Pair of gloves", "Shovel", "Rake", "Watering can", "Lawn mower"], + "Tools": ["Hammer", "Screwdriver", "Wrench", "Saw", "Pliers;plier;Pair of pliers", "Drill"] +} + +DEFAULT_OBJECT_LIST = sum([d for d in DEFAULT_OBJECT_DICT.values()], []) +# random.seed(42) +# DEFAULT_OBJECT_LIST = random.sample(DEFAULT_OBJECT_LIST, k=5) + +SUBSET_OBJECT_LIST = [DEFAULT_OBJECT_LIST[i] for i in [1,11,21,31,41,51,61,71,81,91]] +INITIAL_STR = "Questions:\n" # must be changed in the dataset as well + +class TwentyQuestionsEnv(): + def __init__( + self, + # word_list, + max_conversation_length: int=20, + word_list = DEFAULT_OBJECT_LIST + ): + self.word_list = word_list + self.word_list =[ list(map(lambda x: x.lower(), word.split(";"))) for word in self.word_list] + self.max_conversation_length = max_conversation_length + + self.random = random.Random(None) + self.count = 0 + self.curr_word = None + self.history = '' + self.done = True + + def is_correct(self, question): + #check for the last word + # cut out punctuations at the end + while len(question) > 0 and not question[-1].isalpha(): + question = question[:-1] + + if len(question) == 0: + return False + guess = question.split(" ")[-1].lower() + return guess in self.curr_word + + def _step(self, question, answer): + if answer.strip() == 'yes': + answer = 'Yes.' + elif answer.strip() == 'no': + answer = 'No.' + if self.done: + return None + self.count+=1 + self.history += question + ' ' + answer + '\n' + + # trajectory = create_trajectory_from_history(self.curr_word, text_history + (answer_text,), self.max_conversation_length) + # if self.count == self.max_conversation_length: + # print("The word was", self.curr_word[0]) + done = (answer.replace('.', '').lower() == 'yes') and self.is_correct(question) + reward = -1 + if done: + reward = 0 + self.done = done or self.count == self.max_conversation_length + return self.history, reward, self.done + + def step(self, question): + if self.done: + return None + assert self.curr_word is not None, "call env.reset() first." + answer = self.generate_answer(question) + return self._step(question, answer) + + # return trajectory.text_history, trajectory.reward[-2], trajectory.done + + def reset(self, idx : Optional[int]=None): + self.count = 0 + # if self.curr_word is not None: + # print("The word was ", self.curr_word) + # print("Next word...") + if idx is not None: + self.curr_word = self.word_list[idx] + else: + self.curr_word = self.random.choice(self.word_list) + # print(self.word_list) + # exit() + self.history = INITIAL_STR + self.done = False + return INITIAL_STR + # return (Text(INITIAL_STR, is_action=False),) + + def copy(self): + return TwentyQuestionsEnv( + oracle=self.oracle, + word_list=self.word_list, + max_conversation_length=self.max_conversation_length, + ) + + +class BatchedTwentyQuestionsEnv(): + + def __init__( + self, + max_conversation_length: int=20, + bsize: int=32, + # device = None, + word_list = DEFAULT_OBJECT_LIST + ): + word_list = DEFAULT_OBJECT_LIST if word_list is None else word_list + self.env_list = [TwentyQuestionsEnv(max_conversation_length, word_list = word_list) for _ in range(bsize)] + self.bsize = bsize + + def generate_answers(self, questions): + curr_words = [env.curr_word[0].lower() for env in self.env_list] + inputs = [f"The object is {curr_word}." + question for curr_word, question in zip(curr_words, questions)] + encoder_ids = self.tokenizer(inputs ,padding=True, return_tensors='pt').to(self.model.device) + return self.tokenizer.batch_decode(self.model.generate(input_ids=encoder_ids['input_ids'], attention_mask=encoder_ids['attention_mask'],\ + max_new_tokens=16), skip_special_tokens= True) + + def reset(self, idx: Optional[int] = None): + return [env.reset(idx) for env in self.env_list] + + def step(self, questions): + answers = self.generate_answers(questions) + # print("Step once!") + with concurrent.futures.ThreadPoolExecutor() as executor: + jobs = [executor.submit(env._step, q, a) for env, q, a in zip(self.env_list, questions, answers)] + results = [job.result() for job in jobs] + return results \ No newline at end of file diff --git a/wandb/debug-internal.log b/wandb/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..53f1f5b66b5caf082f34bec47dd79b998bf0f2e5 --- /dev/null +++ b/wandb/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T23:00:58.141834544+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T23:00:58.173649681+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T23:00:58.638565435+08:00","level":"INFO","msg":"stream: created new stream","id":"ao2fiuap"} +{"time":"2025-09-16T23:00:58.638719575+08:00","level":"INFO","msg":"stream: started","id":"ao2fiuap"} +{"time":"2025-09-16T23:00:58.63887401+08:00","level":"INFO","msg":"writer: started","stream_id":"ao2fiuap"} +{"time":"2025-09-16T23:00:58.639098632+08:00","level":"INFO","msg":"handler: started","stream_id":"ao2fiuap"} +{"time":"2025-09-16T23:00:58.639062744+08:00","level":"INFO","msg":"sender: started","stream_id":"ao2fiuap"} diff --git a/wandb/debug.log b/wandb/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..16e4958b9a7c1c2ec9c5e06bf776b9021ebf182d --- /dev/null +++ b/wandb/debug.log @@ -0,0 +1,22 @@ +2025-09-16 23:00:57,903 INFO MainThread:185929 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_setup.py:_flush():81] Configure stats pid to 185929 +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_230057-ao2fiuap/logs/debug.log +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_230057-ao2fiuap/logs/debug-internal.log +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_init.py:init():813] calling init triggers +2025-09-16 23:00:57,904 INFO MainThread:185929 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 23:00:57,905 INFO MainThread:185929 [wandb_init.py:init():854] starting backend +2025-09-16 23:00:58,127 INFO MainThread:185929 [wandb_init.py:init():857] sending inform_init request +2025-09-16 23:00:58,132 INFO MainThread:185929 [wandb_init.py:init():865] backend started and connected +2025-09-16 23:00:58,135 INFO MainThread:185929 [wandb_init.py:init():936] updated telemetry +2025-09-16 23:00:58,149 INFO MainThread:185929 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 23:00:58,987 INFO MainThread:185929 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 23:00:59,148 INFO MainThread:185929 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 23:00:59,148 INFO MainThread:185929 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 23:00:59,148 INFO MainThread:185929 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 23:00:59,148 INFO MainThread:185929 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 23:00:59,151 INFO MainThread:185929 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 23:02:21,240 INFO MainThread:185929 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250914_015126-uv3mxna5/files/output.log b/wandb/run-20250914_015126-uv3mxna5/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_015126-uv3mxna5/files/requirements.txt b/wandb/run-20250914_015126-uv3mxna5/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..233fe1c5f2b48d9c2db2be552c464c7ac53511f0 --- /dev/null +++ b/wandb/run-20250914_015126-uv3mxna5/files/requirements.txt @@ -0,0 +1,99 @@ +archer==0.1.0 +pip==25.2 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +psutil==7.0.0 +nvidia-cufile-cu12==1.13.1.3 +charset-normalizer==3.4.3 +lightning==2.5.5 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +ninja==1.13.0 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +attrs==25.3.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +py-cpuinfo==9.0.0 +nvidia-cusolver-cu12==11.7.3.90 +Jinja2==3.1.6 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +nvidia-nvtx-cu12==12.8.90 +GitPython==3.1.45 +click==8.2.1 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +safetensors==0.6.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +hf-xet==1.1.10 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_015126-uv3mxna5/files/wandb-metadata.json b/wandb/run-20250914_015126-uv3mxna5/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..59ad591423316d5eeb99c10ba8c4c7ed1d61b124 --- /dev/null +++ b/wandb/run-20250914_015126-uv3mxna5/files/wandb-metadata.json @@ -0,0 +1,53 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T17:51:26.688659Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=32", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3454672388096" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "swdx3pzr4n3ibfgn11mng6euu0bahshw" +} \ No newline at end of file diff --git a/wandb/run-20250914_015126-uv3mxna5/logs/debug-core.log b/wandb/run-20250914_015126-uv3mxna5/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..4c3034df28e926fd8a29aba51d955b70fe5e073a --- /dev/null +++ b/wandb/run-20250914_015126-uv3mxna5/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T01:51:26.954919787+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp9ry4pp1h/port-2979525.txt","pid":2979525,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T01:51:26.955168197+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T01:51:26.956172114+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2979525} +{"time":"2025-09-14T01:51:26.956166515+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2979525-2986911-2612594414/socket","Net":"unix"}} +{"time":"2025-09-14T01:51:27.144249317+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T01:51:27.173942723+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"uv3mxna5","id":"1(@)"} +{"time":"2025-09-14T01:51:27.669926106+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"uv3mxna5","id":"1(@)"} +{"time":"2025-09-14T01:53:03.661418154+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_015126-uv3mxna5/logs/debug-internal.log b/wandb/run-20250914_015126-uv3mxna5/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..ff4500ed2f4e0a6bcd0fb7cead5b817c4121b7b0 --- /dev/null +++ b/wandb/run-20250914_015126-uv3mxna5/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T01:51:27.1740897+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T01:51:27.218209764+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T01:51:27.669836334+08:00","level":"INFO","msg":"stream: created new stream","id":"uv3mxna5"} +{"time":"2025-09-14T01:51:27.669911752+08:00","level":"INFO","msg":"stream: started","id":"uv3mxna5"} +{"time":"2025-09-14T01:51:27.669933883+08:00","level":"INFO","msg":"writer: started","stream_id":"uv3mxna5"} +{"time":"2025-09-14T01:51:27.669951233+08:00","level":"INFO","msg":"sender: started","stream_id":"uv3mxna5"} +{"time":"2025-09-14T01:51:27.670046072+08:00","level":"INFO","msg":"handler: started","stream_id":"uv3mxna5"} diff --git a/wandb/run-20250914_015126-uv3mxna5/logs/debug.log b/wandb/run-20250914_015126-uv3mxna5/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e86c6250dce78d5f33e39918f504dbcff60ca099 --- /dev/null +++ b/wandb/run-20250914_015126-uv3mxna5/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 01:51:26,694 INFO MainThread:2979525 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 01:51:26,695 INFO MainThread:2979525 [wandb_setup.py:_flush():81] Configure stats pid to 2979525 +2025-09-14 01:51:26,695 INFO MainThread:2979525 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 01:51:26,695 INFO MainThread:2979525 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 01:51:26,695 INFO MainThread:2979525 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 01:51:26,696 INFO MainThread:2979525 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_015126-uv3mxna5/logs/debug.log +2025-09-14 01:51:26,696 INFO MainThread:2979525 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_015126-uv3mxna5/logs/debug-internal.log +2025-09-14 01:51:26,697 INFO MainThread:2979525 [wandb_init.py:init():813] calling init triggers +2025-09-14 01:51:26,697 INFO MainThread:2979525 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 01:51:26,697 INFO MainThread:2979525 [wandb_init.py:init():854] starting backend +2025-09-14 01:51:27,145 INFO MainThread:2979525 [wandb_init.py:init():857] sending inform_init request +2025-09-14 01:51:27,170 INFO MainThread:2979525 [wandb_init.py:init():865] backend started and connected +2025-09-14 01:51:27,176 INFO MainThread:2979525 [wandb_init.py:init():936] updated telemetry +2025-09-14 01:51:27,203 INFO MainThread:2979525 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 01:51:28,009 INFO MainThread:2979525 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 01:51:28,543 INFO MainThread:2979525 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 01:51:28,545 INFO MainThread:2979525 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 01:51:28,546 INFO MainThread:2979525 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 01:51:28,546 INFO MainThread:2979525 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 01:51:28,577 INFO MainThread:2979525 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_015126-uv3mxna5/run-uv3mxna5.wandb b/wandb/run-20250914_015126-uv3mxna5/run-uv3mxna5.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_015639-yhfx3ktc/files/config.yaml b/wandb/run-20250914_015639-yhfx3ktc/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3aa3afbf43594d45c1ca23f4e0ecbebd69a0f997 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/files/config.yaml @@ -0,0 +1,84 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + 65ypv1qtzwxxxt44f89vzf2tnsbx3g27: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=64 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.val_check_interval=10000 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3458710171648" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T17:56:39.014301Z" + writerId: 65ypv1qtzwxxxt44f89vzf2tnsbx3g27 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_015639-yhfx3ktc/files/output.log b/wandb/run-20250914_015639-yhfx3ktc/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..51cfb21926742c48ecda010752ddb72288af70b2 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/files/output.log @@ -0,0 +1,100 @@ +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 125, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 156, in read_data + outcome, history_length = get_game_outcome( + File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 232, in get_game_outcome + if item["role"] == "listener": + File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 211, in is_correct_prediction + return PSTEMMER.stem(word[:-1]) + File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 210, in custom_stem + if word.endswith("y") and len(word) > 3: +NameError: name 'PSTEMMER' is not defined +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 125, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 156, in read_data +[rank0]: outcome, history_length = get_game_outcome( +[rank0]: File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 232, in get_game_outcome +[rank0]: if item["role"] == "listener": +[rank0]: File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 211, in is_correct_prediction +[rank0]: return PSTEMMER.stem(word[:-1]) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 210, in custom_stem +[rank0]: if word.endswith("y") and len(word) > 3: +[rank0]: NameError: name 'PSTEMMER' is not defined diff --git a/wandb/run-20250914_015639-yhfx3ktc/files/requirements.txt b/wandb/run-20250914_015639-yhfx3ktc/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..233fe1c5f2b48d9c2db2be552c464c7ac53511f0 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/files/requirements.txt @@ -0,0 +1,99 @@ +archer==0.1.0 +pip==25.2 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +psutil==7.0.0 +nvidia-cufile-cu12==1.13.1.3 +charset-normalizer==3.4.3 +lightning==2.5.5 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +ninja==1.13.0 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +attrs==25.3.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +py-cpuinfo==9.0.0 +nvidia-cusolver-cu12==11.7.3.90 +Jinja2==3.1.6 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +nvidia-nvtx-cu12==12.8.90 +GitPython==3.1.45 +click==8.2.1 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +safetensors==0.6.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +hf-xet==1.1.10 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_015639-yhfx3ktc/files/wandb-metadata.json b/wandb/run-20250914_015639-yhfx3ktc/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6d6ebc746fe314ee65460ebe7c50901d137a9706 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/files/wandb-metadata.json @@ -0,0 +1,53 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T17:56:39.014301Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3458710171648" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "65ypv1qtzwxxxt44f89vzf2tnsbx3g27" +} \ No newline at end of file diff --git a/wandb/run-20250914_015639-yhfx3ktc/files/wandb-summary.json b/wandb/run-20250914_015639-yhfx3ktc/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..f9e4843abb5d7f611d55c94e912ecc0e50067f04 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":47},"_runtime":47} \ No newline at end of file diff --git a/wandb/run-20250914_015639-yhfx3ktc/logs/debug-core.log b/wandb/run-20250914_015639-yhfx3ktc/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..fc308b195167d1c713c9e12fb3ee058db7f7cd1a --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T01:56:39.060768088+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmph20ishe3/port-2994656.txt","pid":2994656,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T01:56:39.060980154+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T01:56:39.061593543+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2994656} +{"time":"2025-09-14T01:56:39.061587928+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2994656-3006850-3590603342/socket","Net":"unix"}} +{"time":"2025-09-14T01:56:39.250476135+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T01:56:39.273820267+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"yhfx3ktc","id":"1(@)"} +{"time":"2025-09-14T01:56:39.784042789+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"yhfx3ktc","id":"1(@)"} +{"time":"2025-09-14T01:57:27.406243024+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T01:57:27.406587671+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T01:57:27.40678208+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T01:57:27.406634198+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T01:57:27.4072589+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2994656-3006850-3590603342/socket","Net":"unix"}} +{"time":"2025-09-14T01:57:28.951616553+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T01:57:28.951671653+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T01:57:28.951697038+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_015639-yhfx3ktc/logs/debug-internal.log b/wandb/run-20250914_015639-yhfx3ktc/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..da655bf3d73d77479451c512ded06e19c35463c1 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T01:56:39.274134476+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T01:56:39.308522468+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T01:56:39.783928706+08:00","level":"INFO","msg":"stream: created new stream","id":"yhfx3ktc"} +{"time":"2025-09-14T01:56:39.784027362+08:00","level":"INFO","msg":"stream: started","id":"yhfx3ktc"} +{"time":"2025-09-14T01:56:39.784053277+08:00","level":"INFO","msg":"writer: started","stream_id":"yhfx3ktc"} +{"time":"2025-09-14T01:56:39.784072896+08:00","level":"INFO","msg":"handler: started","stream_id":"yhfx3ktc"} +{"time":"2025-09-14T01:56:39.784114892+08:00","level":"INFO","msg":"sender: started","stream_id":"yhfx3ktc"} +{"time":"2025-09-14T01:57:27.406437558+08:00","level":"INFO","msg":"stream: closing","id":"yhfx3ktc"} +{"time":"2025-09-14T01:57:28.68044474+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T01:57:28.949926089+08:00","level":"INFO","msg":"handler: closed","stream_id":"yhfx3ktc"} +{"time":"2025-09-14T01:57:28.950701948+08:00","level":"INFO","msg":"sender: closed","stream_id":"yhfx3ktc"} +{"time":"2025-09-14T01:57:28.950734117+08:00","level":"INFO","msg":"stream: closed","id":"yhfx3ktc"} diff --git a/wandb/run-20250914_015639-yhfx3ktc/logs/debug.log b/wandb/run-20250914_015639-yhfx3ktc/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..9e101777e2152c79d64b0114e0baebeb62285696 --- /dev/null +++ b/wandb/run-20250914_015639-yhfx3ktc/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 01:56:39,023 INFO MainThread:2994656 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 01:56:39,024 INFO MainThread:2994656 [wandb_setup.py:_flush():81] Configure stats pid to 2994656 +2025-09-14 01:56:39,024 INFO MainThread:2994656 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 01:56:39,024 INFO MainThread:2994656 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 01:56:39,024 INFO MainThread:2994656 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 01:56:39,025 INFO MainThread:2994656 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_015639-yhfx3ktc/logs/debug.log +2025-09-14 01:56:39,025 INFO MainThread:2994656 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_015639-yhfx3ktc/logs/debug-internal.log +2025-09-14 01:56:39,026 INFO MainThread:2994656 [wandb_init.py:init():813] calling init triggers +2025-09-14 01:56:39,026 INFO MainThread:2994656 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 01:56:39,026 INFO MainThread:2994656 [wandb_init.py:init():854] starting backend +2025-09-14 01:56:39,249 INFO MainThread:2994656 [wandb_init.py:init():857] sending inform_init request +2025-09-14 01:56:39,260 INFO MainThread:2994656 [wandb_init.py:init():865] backend started and connected +2025-09-14 01:56:39,271 INFO MainThread:2994656 [wandb_init.py:init():936] updated telemetry +2025-09-14 01:56:39,300 INFO MainThread:2994656 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 01:56:40,060 INFO MainThread:2994656 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 01:56:40,332 INFO MainThread:2994656 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 01:56:40,333 INFO MainThread:2994656 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 01:56:40,333 INFO MainThread:2994656 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 01:56:40,334 INFO MainThread:2994656 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 01:56:40,338 INFO MainThread:2994656 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 01:57:27,406 INFO wandb-AsyncioManager-main:2994656 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 01:57:27,407 INFO wandb-AsyncioManager-main:2994656 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_015639-yhfx3ktc/run-yhfx3ktc.wandb b/wandb/run-20250914_015639-yhfx3ktc/run-yhfx3ktc.wandb new file mode 100644 index 0000000000000000000000000000000000000000..3a2b76deb886a174b67fb8e90e7a429159961d69 Binary files /dev/null and b/wandb/run-20250914_015639-yhfx3ktc/run-yhfx3ktc.wandb differ diff --git a/wandb/run-20250914_020205-omtiahcp/files/config.yaml b/wandb/run-20250914_020205-omtiahcp/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e6104353b22a83d9ee72be9923da06ebf9c6abad --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/files/config.yaml @@ -0,0 +1,85 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + 3x3r0y3j6elbxx2nguycn1hxqpdais0n: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=64 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.val_check_interval=10000 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3481643978752" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T18:02:05.736654Z" + writerId: 3x3r0y3j6elbxx2nguycn1hxqpdais0n + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_020205-omtiahcp/files/output.log b/wandb/run-20250914_020205-omtiahcp/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ed5b47f7688d4b5031e6eb3585aac9bd3bbe6346 --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/files/output.log @@ -0,0 +1,92 @@ +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 126, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 135, in read_data + from rsa_game import get_game_outcome, randomly_convert_game_history_to_query + File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 5, in + from nltk.stem import PorterStemmer +ModuleNotFoundError: No module named 'nltk' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 126, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 135, in read_data +[rank0]: from rsa_game import get_game_outcome, randomly_convert_game_history_to_query +[rank0]: File "/home/jiashuo/codes/OfflineArcher/rsa_game.py", line 5, in +[rank0]: from nltk.stem import PorterStemmer +[rank0]: ModuleNotFoundError: No module named 'nltk' diff --git a/wandb/run-20250914_020205-omtiahcp/files/requirements.txt b/wandb/run-20250914_020205-omtiahcp/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..233fe1c5f2b48d9c2db2be552c464c7ac53511f0 --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/files/requirements.txt @@ -0,0 +1,99 @@ +archer==0.1.0 +pip==25.2 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +psutil==7.0.0 +nvidia-cufile-cu12==1.13.1.3 +charset-normalizer==3.4.3 +lightning==2.5.5 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +ninja==1.13.0 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +attrs==25.3.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +py-cpuinfo==9.0.0 +nvidia-cusolver-cu12==11.7.3.90 +Jinja2==3.1.6 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +nvidia-nvtx-cu12==12.8.90 +GitPython==3.1.45 +click==8.2.1 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +safetensors==0.6.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +hf-xet==1.1.10 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_020205-omtiahcp/files/wandb-metadata.json b/wandb/run-20250914_020205-omtiahcp/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6643f7fe142ba2bd015d10b384133fefa57d2e40 --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T18:02:05.736654Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3481643978752" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "3x3r0y3j6elbxx2nguycn1hxqpdais0n" +} \ No newline at end of file diff --git a/wandb/run-20250914_020205-omtiahcp/files/wandb-summary.json b/wandb/run-20250914_020205-omtiahcp/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..fed6cb7d2bad8189db5a1a804105a27f56f9ac60 --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":21,"_wandb":{"runtime":21}} \ No newline at end of file diff --git a/wandb/run-20250914_020205-omtiahcp/logs/debug-core.log b/wandb/run-20250914_020205-omtiahcp/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..193567a7670a637c33b9ad31d732f33b65a0dcc0 --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T02:02:05.804288588+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp5i5bnkxl/port-3017092.txt","pid":3017092,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T02:02:05.804494832+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T02:02:05.805223929+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3017092} +{"time":"2025-09-14T02:02:05.805234311+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3017092-3025175-1151521267/socket","Net":"unix"}} +{"time":"2025-09-14T02:02:05.992683363+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T02:02:06.012893944+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"omtiahcp","id":"1(@)"} +{"time":"2025-09-14T02:02:06.46073686+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"omtiahcp","id":"1(@)"} +{"time":"2025-09-14T02:02:28.643923683+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T02:02:28.644277844+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T02:02:28.644267091+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T02:02:28.644505889+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T02:02:28.644883915+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3017092-3025175-1151521267/socket","Net":"unix"}} +{"time":"2025-09-14T02:02:30.248988734+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T02:02:30.249055745+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T02:02:30.24907562+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_020205-omtiahcp/logs/debug-internal.log b/wandb/run-20250914_020205-omtiahcp/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..462c54989df423619f0d5fc69904e15365981ace --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T02:02:06.013034885+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T02:02:06.032139592+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T02:02:06.460619393+08:00","level":"INFO","msg":"stream: created new stream","id":"omtiahcp"} +{"time":"2025-09-14T02:02:06.460722006+08:00","level":"INFO","msg":"stream: started","id":"omtiahcp"} +{"time":"2025-09-14T02:02:06.460756201+08:00","level":"INFO","msg":"writer: started","stream_id":"omtiahcp"} +{"time":"2025-09-14T02:02:06.460770286+08:00","level":"INFO","msg":"handler: started","stream_id":"omtiahcp"} +{"time":"2025-09-14T02:02:06.460869627+08:00","level":"INFO","msg":"sender: started","stream_id":"omtiahcp"} +{"time":"2025-09-14T02:02:28.64411066+08:00","level":"INFO","msg":"stream: closing","id":"omtiahcp"} +{"time":"2025-09-14T02:02:29.949747088+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T02:02:30.246552435+08:00","level":"INFO","msg":"handler: closed","stream_id":"omtiahcp"} +{"time":"2025-09-14T02:02:30.247497693+08:00","level":"INFO","msg":"sender: closed","stream_id":"omtiahcp"} +{"time":"2025-09-14T02:02:30.247557187+08:00","level":"INFO","msg":"stream: closed","id":"omtiahcp"} diff --git a/wandb/run-20250914_020205-omtiahcp/logs/debug.log b/wandb/run-20250914_020205-omtiahcp/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a6651bb09defb3808494e1b6f9ee8bf9edbe0b45 --- /dev/null +++ b/wandb/run-20250914_020205-omtiahcp/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 02:02:05,743 INFO MainThread:3017092 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 02:02:05,743 INFO MainThread:3017092 [wandb_setup.py:_flush():81] Configure stats pid to 3017092 +2025-09-14 02:02:05,744 INFO MainThread:3017092 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 02:02:05,744 INFO MainThread:3017092 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 02:02:05,744 INFO MainThread:3017092 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 02:02:05,745 INFO MainThread:3017092 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_020205-omtiahcp/logs/debug.log +2025-09-14 02:02:05,745 INFO MainThread:3017092 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_020205-omtiahcp/logs/debug-internal.log +2025-09-14 02:02:05,745 INFO MainThread:3017092 [wandb_init.py:init():813] calling init triggers +2025-09-14 02:02:05,746 INFO MainThread:3017092 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 02:02:05,746 INFO MainThread:3017092 [wandb_init.py:init():854] starting backend +2025-09-14 02:02:05,993 INFO MainThread:3017092 [wandb_init.py:init():857] sending inform_init request +2025-09-14 02:02:06,004 INFO MainThread:3017092 [wandb_init.py:init():865] backend started and connected +2025-09-14 02:02:06,010 INFO MainThread:3017092 [wandb_init.py:init():936] updated telemetry +2025-09-14 02:02:06,037 INFO MainThread:3017092 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 02:02:06,744 INFO MainThread:3017092 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 02:02:07,071 INFO MainThread:3017092 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 02:02:07,072 INFO MainThread:3017092 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 02:02:07,073 INFO MainThread:3017092 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 02:02:07,073 INFO MainThread:3017092 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 02:02:07,079 INFO MainThread:3017092 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 02:02:28,644 INFO wandb-AsyncioManager-main:3017092 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 02:02:28,644 INFO wandb-AsyncioManager-main:3017092 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_020205-omtiahcp/run-omtiahcp.wandb b/wandb/run-20250914_020205-omtiahcp/run-omtiahcp.wandb new file mode 100644 index 0000000000000000000000000000000000000000..0e16b65fc3746eee11eb2a0bfcb43efb8d925c6d Binary files /dev/null and b/wandb/run-20250914_020205-omtiahcp/run-omtiahcp.wandb differ diff --git a/wandb/run-20250914_020524-049c4zvx/files/output.log b/wandb/run-20250914_020524-049c4zvx/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_020524-049c4zvx/files/requirements.txt b/wandb/run-20250914_020524-049c4zvx/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..38bf7b3dbee06392e7dbb05e4b0bbf7b34b97de9 --- /dev/null +++ b/wandb/run-20250914_020524-049c4zvx/files/requirements.txt @@ -0,0 +1,101 @@ +archer==0.1.0 +pip==25.2 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +psutil==7.0.0 +nvidia-cufile-cu12==1.13.1.3 +charset-normalizer==3.4.3 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +ninja==1.13.0 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +attrs==25.3.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +py-cpuinfo==9.0.0 +nvidia-cusolver-cu12==11.7.3.90 +Jinja2==3.1.6 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +nvidia-nvtx-cu12==12.8.90 +GitPython==3.1.45 +click==8.2.1 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +safetensors==0.6.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +hf-xet==1.1.10 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_020524-049c4zvx/files/wandb-metadata.json b/wandb/run-20250914_020524-049c4zvx/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7a9498e130499d213449b6d44b666faa0dd85574 --- /dev/null +++ b/wandb/run-20250914_020524-049c4zvx/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T18:05:24.539990Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3490523860992" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "p3cvj74qu51zadwqnzz2pzbmfj8cibnx" +} \ No newline at end of file diff --git a/wandb/run-20250914_020524-049c4zvx/logs/debug-core.log b/wandb/run-20250914_020524-049c4zvx/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..4d6ac84a5c10e47998fdfc12f44b3192b0efca41 --- /dev/null +++ b/wandb/run-20250914_020524-049c4zvx/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T02:05:24.608970174+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpn4d02pyh/port-3028499.txt","pid":3028499,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T02:05:24.609247293+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T02:05:24.609972851+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3028499} +{"time":"2025-09-14T02:05:24.609978246+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3028499-3038318-998366716/socket","Net":"unix"}} +{"time":"2025-09-14T02:05:24.798618436+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T02:05:24.822142842+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"049c4zvx","id":"1(@)"} +{"time":"2025-09-14T02:05:25.310624102+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"049c4zvx","id":"1(@)"} +{"time":"2025-09-14T02:08:26.387758357+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_020524-049c4zvx/logs/debug-internal.log b/wandb/run-20250914_020524-049c4zvx/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..7260d3ac35fdea19aff4ed61946b912311cf7b68 --- /dev/null +++ b/wandb/run-20250914_020524-049c4zvx/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T02:05:24.822486784+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T02:05:24.862009562+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T02:05:25.310539722+08:00","level":"INFO","msg":"stream: created new stream","id":"049c4zvx"} +{"time":"2025-09-14T02:05:25.310610292+08:00","level":"INFO","msg":"stream: started","id":"049c4zvx"} +{"time":"2025-09-14T02:05:25.310638725+08:00","level":"INFO","msg":"writer: started","stream_id":"049c4zvx"} +{"time":"2025-09-14T02:05:25.31066039+08:00","level":"INFO","msg":"sender: started","stream_id":"049c4zvx"} +{"time":"2025-09-14T02:05:25.310683213+08:00","level":"INFO","msg":"handler: started","stream_id":"049c4zvx"} diff --git a/wandb/run-20250914_020524-049c4zvx/logs/debug.log b/wandb/run-20250914_020524-049c4zvx/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..576a4954b7408b4da038a011940947a572588b65 --- /dev/null +++ b/wandb/run-20250914_020524-049c4zvx/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 02:05:24,547 INFO MainThread:3028499 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 02:05:24,547 INFO MainThread:3028499 [wandb_setup.py:_flush():81] Configure stats pid to 3028499 +2025-09-14 02:05:24,547 INFO MainThread:3028499 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 02:05:24,548 INFO MainThread:3028499 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 02:05:24,548 INFO MainThread:3028499 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 02:05:24,548 INFO MainThread:3028499 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_020524-049c4zvx/logs/debug.log +2025-09-14 02:05:24,549 INFO MainThread:3028499 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_020524-049c4zvx/logs/debug-internal.log +2025-09-14 02:05:24,549 INFO MainThread:3028499 [wandb_init.py:init():813] calling init triggers +2025-09-14 02:05:24,549 INFO MainThread:3028499 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 02:05:24,550 INFO MainThread:3028499 [wandb_init.py:init():854] starting backend +2025-09-14 02:05:24,799 INFO MainThread:3028499 [wandb_init.py:init():857] sending inform_init request +2025-09-14 02:05:24,810 INFO MainThread:3028499 [wandb_init.py:init():865] backend started and connected +2025-09-14 02:05:24,819 INFO MainThread:3028499 [wandb_init.py:init():936] updated telemetry +2025-09-14 02:05:24,844 INFO MainThread:3028499 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 02:05:25,583 INFO MainThread:3028499 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 02:05:25,894 INFO MainThread:3028499 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 02:05:25,895 INFO MainThread:3028499 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 02:05:25,895 INFO MainThread:3028499 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 02:05:25,895 INFO MainThread:3028499 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 02:05:25,901 INFO MainThread:3028499 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_020524-049c4zvx/run-049c4zvx.wandb b/wandb/run-20250914_020524-049c4zvx/run-049c4zvx.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_021216-qsurteoc/files/output.log b/wandb/run-20250914_021216-qsurteoc/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_021216-qsurteoc/files/requirements.txt b/wandb/run-20250914_021216-qsurteoc/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..38bf7b3dbee06392e7dbb05e4b0bbf7b34b97de9 --- /dev/null +++ b/wandb/run-20250914_021216-qsurteoc/files/requirements.txt @@ -0,0 +1,101 @@ +archer==0.1.0 +pip==25.2 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +psutil==7.0.0 +nvidia-cufile-cu12==1.13.1.3 +charset-normalizer==3.4.3 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +ninja==1.13.0 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +attrs==25.3.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +py-cpuinfo==9.0.0 +nvidia-cusolver-cu12==11.7.3.90 +Jinja2==3.1.6 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +nvidia-nvtx-cu12==12.8.90 +GitPython==3.1.45 +click==8.2.1 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +safetensors==0.6.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +hf-xet==1.1.10 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_021216-qsurteoc/files/wandb-metadata.json b/wandb/run-20250914_021216-qsurteoc/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6c377f71ee51b281c8886c2ee3bb71ff1c906e04 --- /dev/null +++ b/wandb/run-20250914_021216-qsurteoc/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T18:12:16.751247Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3498792079360" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "qusqd8833v0qk5xx8gjpz6hs38r48wsg" +} \ No newline at end of file diff --git a/wandb/run-20250914_021216-qsurteoc/logs/debug-core.log b/wandb/run-20250914_021216-qsurteoc/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..c2da82868660cf196cdc003d2b13abf0fedc17ee --- /dev/null +++ b/wandb/run-20250914_021216-qsurteoc/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T02:12:16.82743332+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpmyihxtbc/port-3061676.txt","pid":3061676,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T02:12:16.827703696+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T02:12:16.829798944+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3061676} +{"time":"2025-09-14T02:12:16.829747196+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3061676-3070459-973020177/socket","Net":"unix"}} +{"time":"2025-09-14T02:12:17.016516183+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T02:12:17.033455127+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"qsurteoc","id":"1(@)"} +{"time":"2025-09-14T02:12:17.523333281+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"qsurteoc","id":"1(@)"} +{"time":"2025-09-14T02:12:31.184147492+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_021216-qsurteoc/logs/debug-internal.log b/wandb/run-20250914_021216-qsurteoc/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..bbe7e9d0b8a7a628b2fce13803fdb70109c46cc7 --- /dev/null +++ b/wandb/run-20250914_021216-qsurteoc/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T02:12:17.033696372+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T02:12:17.079151427+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T02:12:17.52324+08:00","level":"INFO","msg":"stream: created new stream","id":"qsurteoc"} +{"time":"2025-09-14T02:12:17.52331852+08:00","level":"INFO","msg":"stream: started","id":"qsurteoc"} +{"time":"2025-09-14T02:12:17.523325493+08:00","level":"INFO","msg":"handler: started","stream_id":"qsurteoc"} +{"time":"2025-09-14T02:12:17.523368575+08:00","level":"INFO","msg":"writer: started","stream_id":"qsurteoc"} +{"time":"2025-09-14T02:12:17.523350586+08:00","level":"INFO","msg":"sender: started","stream_id":"qsurteoc"} diff --git a/wandb/run-20250914_021216-qsurteoc/logs/debug.log b/wandb/run-20250914_021216-qsurteoc/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..1b426ff02e1b473019a27badcc0cc5a341952df1 --- /dev/null +++ b/wandb/run-20250914_021216-qsurteoc/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 02:12:16,761 INFO MainThread:3061676 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 02:12:16,761 INFO MainThread:3061676 [wandb_setup.py:_flush():81] Configure stats pid to 3061676 +2025-09-14 02:12:16,762 INFO MainThread:3061676 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 02:12:16,762 INFO MainThread:3061676 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 02:12:16,762 INFO MainThread:3061676 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 02:12:16,763 INFO MainThread:3061676 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_021216-qsurteoc/logs/debug.log +2025-09-14 02:12:16,763 INFO MainThread:3061676 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_021216-qsurteoc/logs/debug-internal.log +2025-09-14 02:12:16,764 INFO MainThread:3061676 [wandb_init.py:init():813] calling init triggers +2025-09-14 02:12:16,764 INFO MainThread:3061676 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 02:12:16,764 INFO MainThread:3061676 [wandb_init.py:init():854] starting backend +2025-09-14 02:12:17,017 INFO MainThread:3061676 [wandb_init.py:init():857] sending inform_init request +2025-09-14 02:12:17,029 INFO MainThread:3061676 [wandb_init.py:init():865] backend started and connected +2025-09-14 02:12:17,041 INFO MainThread:3061676 [wandb_init.py:init():936] updated telemetry +2025-09-14 02:12:17,080 INFO MainThread:3061676 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 02:12:17,886 INFO MainThread:3061676 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 02:12:18,247 INFO MainThread:3061676 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 02:12:18,248 INFO MainThread:3061676 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 02:12:18,248 INFO MainThread:3061676 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 02:12:18,248 INFO MainThread:3061676 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 02:12:18,254 INFO MainThread:3061676 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_021216-qsurteoc/run-qsurteoc.wandb b/wandb/run-20250914_021216-qsurteoc/run-qsurteoc.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_021657-ctlp266r/files/config.yaml b/wandb/run-20250914_021657-ctlp266r/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..aa17efd8610a1fbc45233cea031d08930b664df9 --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/files/config.yaml @@ -0,0 +1,85 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + w05bl7ls595kdpg5fjwu8bran97xczxh: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=64 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.val_check_interval=10000 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3503879610368" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T18:16:57.054768Z" + writerId: w05bl7ls595kdpg5fjwu8bran97xczxh + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_021657-ctlp266r/files/output.log b/wandb/run-20250914_021657-ctlp266r/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ad5d0c01703efb1a75332a42f060c7d484368d68 --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/files/output.log @@ -0,0 +1,91 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run + self.strategy.setup(self) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup + self.init_deepspeed() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed + self._initialize_deepspeed_train(self.model) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 492, in _initialize_deepspeed_train + ) = self._init_optimizers() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 468, in _init_optimizers + raise MisconfigurationException( +lightning.fabric.utilities.exceptions.MisconfigurationException: DeepSpeed currently only supports single optimizer, single optional scheduler. +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run +[rank0]: self.strategy.setup(self) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup +[rank0]: self.init_deepspeed() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed +[rank0]: self._initialize_deepspeed_train(self.model) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 492, in _initialize_deepspeed_train +[rank0]: ) = self._init_optimizers() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 468, in _init_optimizers +[rank0]: raise MisconfigurationException( +[rank0]: lightning.fabric.utilities.exceptions.MisconfigurationException: DeepSpeed currently only supports single optimizer, single optional scheduler. diff --git a/wandb/run-20250914_021657-ctlp266r/files/requirements.txt b/wandb/run-20250914_021657-ctlp266r/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..38bf7b3dbee06392e7dbb05e4b0bbf7b34b97de9 --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/files/requirements.txt @@ -0,0 +1,101 @@ +archer==0.1.0 +pip==25.2 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +psutil==7.0.0 +nvidia-cufile-cu12==1.13.1.3 +charset-normalizer==3.4.3 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +ninja==1.13.0 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +attrs==25.3.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +py-cpuinfo==9.0.0 +nvidia-cusolver-cu12==11.7.3.90 +Jinja2==3.1.6 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +nvidia-nvtx-cu12==12.8.90 +GitPython==3.1.45 +click==8.2.1 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +safetensors==0.6.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +hf-xet==1.1.10 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_021657-ctlp266r/files/wandb-metadata.json b/wandb/run-20250914_021657-ctlp266r/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..de21d1dd7fd5303d92562253aef539913635af6a --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T18:16:57.054768Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3503879610368" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "w05bl7ls595kdpg5fjwu8bran97xczxh" +} \ No newline at end of file diff --git a/wandb/run-20250914_021657-ctlp266r/files/wandb-summary.json b/wandb/run-20250914_021657-ctlp266r/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..195123d50401ed52678311a08959c1acb305969b --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":723,"_wandb":{"runtime":723}} \ No newline at end of file diff --git a/wandb/run-20250914_021657-ctlp266r/logs/debug-core.log b/wandb/run-20250914_021657-ctlp266r/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..f29a1821a805bdbf041ad237a0005ec866c97ea0 --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T02:16:57.138474098+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp15atc4r5/port-3075766.txt","pid":3075766,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T02:16:57.138686541+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T02:16:57.139609213+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3075766} +{"time":"2025-09-14T02:16:57.139603767+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3075766-3085844-2573945943/socket","Net":"unix"}} +{"time":"2025-09-14T02:16:57.327900886+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T02:16:57.358683743+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ctlp266r","id":"1(@)"} +{"time":"2025-09-14T02:16:57.826953046+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ctlp266r","id":"1(@)"} +{"time":"2025-09-14T02:29:01.834296905+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T02:29:01.835130837+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T02:29:01.835150049+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T02:29:01.835299122+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T02:29:01.836112725+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3075766-3085844-2573945943/socket","Net":"unix"}} +{"time":"2025-09-14T02:29:03.483355412+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T02:29:03.483416826+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T02:29:03.483455383+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_021657-ctlp266r/logs/debug-internal.log b/wandb/run-20250914_021657-ctlp266r/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8eafcfe088e5df3ae4a57c450e88e52514de1b73 --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T02:16:57.358869834+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T02:16:57.377645618+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T02:16:57.826821953+08:00","level":"INFO","msg":"stream: created new stream","id":"ctlp266r"} +{"time":"2025-09-14T02:16:57.826937758+08:00","level":"INFO","msg":"stream: started","id":"ctlp266r"} +{"time":"2025-09-14T02:16:57.826948218+08:00","level":"INFO","msg":"handler: started","stream_id":"ctlp266r"} +{"time":"2025-09-14T02:16:57.826972421+08:00","level":"INFO","msg":"sender: started","stream_id":"ctlp266r"} +{"time":"2025-09-14T02:16:57.82701654+08:00","level":"INFO","msg":"writer: started","stream_id":"ctlp266r"} +{"time":"2025-09-14T02:29:01.835422432+08:00","level":"INFO","msg":"stream: closing","id":"ctlp266r"} +{"time":"2025-09-14T02:29:03.226042434+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T02:29:03.481666817+08:00","level":"INFO","msg":"handler: closed","stream_id":"ctlp266r"} +{"time":"2025-09-14T02:29:03.48229946+08:00","level":"INFO","msg":"sender: closed","stream_id":"ctlp266r"} +{"time":"2025-09-14T02:29:03.482380401+08:00","level":"INFO","msg":"stream: closed","id":"ctlp266r"} diff --git a/wandb/run-20250914_021657-ctlp266r/logs/debug.log b/wandb/run-20250914_021657-ctlp266r/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..764c68a1ed302307ca6424c7af94806f0acb6f02 --- /dev/null +++ b/wandb/run-20250914_021657-ctlp266r/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 02:16:57,067 INFO MainThread:3075766 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 02:16:57,068 INFO MainThread:3075766 [wandb_setup.py:_flush():81] Configure stats pid to 3075766 +2025-09-14 02:16:57,068 INFO MainThread:3075766 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 02:16:57,068 INFO MainThread:3075766 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 02:16:57,069 INFO MainThread:3075766 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 02:16:57,070 INFO MainThread:3075766 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_021657-ctlp266r/logs/debug.log +2025-09-14 02:16:57,070 INFO MainThread:3075766 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_021657-ctlp266r/logs/debug-internal.log +2025-09-14 02:16:57,071 INFO MainThread:3075766 [wandb_init.py:init():813] calling init triggers +2025-09-14 02:16:57,072 INFO MainThread:3075766 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 02:16:57,072 INFO MainThread:3075766 [wandb_init.py:init():854] starting backend +2025-09-14 02:16:57,329 INFO MainThread:3075766 [wandb_init.py:init():857] sending inform_init request +2025-09-14 02:16:57,346 INFO MainThread:3075766 [wandb_init.py:init():865] backend started and connected +2025-09-14 02:16:57,354 INFO MainThread:3075766 [wandb_init.py:init():936] updated telemetry +2025-09-14 02:16:57,380 INFO MainThread:3075766 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 02:16:58,139 INFO MainThread:3075766 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 02:16:58,483 INFO MainThread:3075766 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 02:16:58,483 INFO MainThread:3075766 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 02:16:58,484 INFO MainThread:3075766 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 02:16:58,484 INFO MainThread:3075766 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 02:16:58,490 INFO MainThread:3075766 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 02:29:01,834 INFO wandb-AsyncioManager-main:3075766 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 02:29:01,835 INFO wandb-AsyncioManager-main:3075766 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_021657-ctlp266r/run-ctlp266r.wandb b/wandb/run-20250914_021657-ctlp266r/run-ctlp266r.wandb new file mode 100644 index 0000000000000000000000000000000000000000..c0f4d9013691109d13a653220e17a93dbafbcde6 Binary files /dev/null and b/wandb/run-20250914_021657-ctlp266r/run-ctlp266r.wandb differ diff --git a/wandb/run-20250914_025113-gl43w77w/files/config.yaml b/wandb/run-20250914_025113-gl43w77w/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..92c640f3cf31760212677bab51b809f3a1ab006a --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/files/config.yaml @@ -0,0 +1,85 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + w1xhdsb444hb870i9dp4iu5uxrj20b11: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=64 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.val_check_interval=10000 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533137326080" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T18:51:13.475874Z" + writerId: w1xhdsb444hb870i9dp4iu5uxrj20b11 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_025113-gl43w77w/files/output.log b/wandb/run-20250914_025113-gl43w77w/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ad5d0c01703efb1a75332a42f060c7d484368d68 --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/files/output.log @@ -0,0 +1,91 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run + self.strategy.setup(self) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup + self.init_deepspeed() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed + self._initialize_deepspeed_train(self.model) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 492, in _initialize_deepspeed_train + ) = self._init_optimizers() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 468, in _init_optimizers + raise MisconfigurationException( +lightning.fabric.utilities.exceptions.MisconfigurationException: DeepSpeed currently only supports single optimizer, single optional scheduler. +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run +[rank0]: self.strategy.setup(self) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup +[rank0]: self.init_deepspeed() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed +[rank0]: self._initialize_deepspeed_train(self.model) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 492, in _initialize_deepspeed_train +[rank0]: ) = self._init_optimizers() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 468, in _init_optimizers +[rank0]: raise MisconfigurationException( +[rank0]: lightning.fabric.utilities.exceptions.MisconfigurationException: DeepSpeed currently only supports single optimizer, single optional scheduler. diff --git a/wandb/run-20250914_025113-gl43w77w/files/requirements.txt b/wandb/run-20250914_025113-gl43w77w/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..0f41d7d9bd6a8303fbd36805dd1e61ca08114074 --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/files/requirements.txt @@ -0,0 +1,137 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_025113-gl43w77w/files/wandb-metadata.json b/wandb/run-20250914_025113-gl43w77w/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..0a9fa4580ac10a1ab6f5634cd8536c34048f9fc5 --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T18:51:13.475874Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533137326080" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "w1xhdsb444hb870i9dp4iu5uxrj20b11" +} \ No newline at end of file diff --git a/wandb/run-20250914_025113-gl43w77w/files/wandb-summary.json b/wandb/run-20250914_025113-gl43w77w/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..3e3f43553107b44b26f778db06b55d4df2ed18ff --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":125,"_wandb":{"runtime":125}} \ No newline at end of file diff --git a/wandb/run-20250914_025113-gl43w77w/logs/debug-core.log b/wandb/run-20250914_025113-gl43w77w/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..6d25eefddbcec10ec40e98966a02e99e6abc3ce5 --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T02:51:13.598589213+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmplqmtof4o/port-3226566.txt","pid":3226566,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T02:51:13.599175521+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T02:51:13.599941321+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3226566} +{"time":"2025-09-14T02:51:13.599958342+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3226566-3235699-325672892/socket","Net":"unix"}} +{"time":"2025-09-14T02:51:13.753817459+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T02:51:13.793450928+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"gl43w77w","id":"1(@)"} +{"time":"2025-09-14T02:51:14.282402493+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"gl43w77w","id":"1(@)"} +{"time":"2025-09-14T02:53:19.712229369+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T02:53:19.713133342+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T02:53:19.71330091+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T02:53:19.713283589+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T02:53:19.714130109+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3226566-3235699-325672892/socket","Net":"unix"}} +{"time":"2025-09-14T02:53:20.703375842+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T02:53:20.703468567+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T02:53:20.703491634+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_025113-gl43w77w/logs/debug-internal.log b/wandb/run-20250914_025113-gl43w77w/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..f7dbc91c919683280888a461ba2da6bad03916d2 --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T02:51:13.79365789+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T02:51:13.830755347+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T02:51:14.282302245+08:00","level":"INFO","msg":"stream: created new stream","id":"gl43w77w"} +{"time":"2025-09-14T02:51:14.282386918+08:00","level":"INFO","msg":"stream: started","id":"gl43w77w"} +{"time":"2025-09-14T02:51:14.282424295+08:00","level":"INFO","msg":"handler: started","stream_id":"gl43w77w"} +{"time":"2025-09-14T02:51:14.282436456+08:00","level":"INFO","msg":"writer: started","stream_id":"gl43w77w"} +{"time":"2025-09-14T02:51:14.282476759+08:00","level":"INFO","msg":"sender: started","stream_id":"gl43w77w"} +{"time":"2025-09-14T02:53:19.713174955+08:00","level":"INFO","msg":"stream: closing","id":"gl43w77w"} +{"time":"2025-09-14T02:53:20.450859089+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T02:53:20.701128402+08:00","level":"INFO","msg":"handler: closed","stream_id":"gl43w77w"} +{"time":"2025-09-14T02:53:20.70200733+08:00","level":"INFO","msg":"sender: closed","stream_id":"gl43w77w"} +{"time":"2025-09-14T02:53:20.702067454+08:00","level":"INFO","msg":"stream: closed","id":"gl43w77w"} diff --git a/wandb/run-20250914_025113-gl43w77w/logs/debug.log b/wandb/run-20250914_025113-gl43w77w/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..cfde34d988065b60d70d306be49d5c8e1de3941d --- /dev/null +++ b/wandb/run-20250914_025113-gl43w77w/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 02:51:13,486 INFO MainThread:3226566 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 02:51:13,487 INFO MainThread:3226566 [wandb_setup.py:_flush():81] Configure stats pid to 3226566 +2025-09-14 02:51:13,487 INFO MainThread:3226566 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 02:51:13,487 INFO MainThread:3226566 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 02:51:13,488 INFO MainThread:3226566 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 02:51:13,488 INFO MainThread:3226566 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_025113-gl43w77w/logs/debug.log +2025-09-14 02:51:13,488 INFO MainThread:3226566 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_025113-gl43w77w/logs/debug-internal.log +2025-09-14 02:51:13,489 INFO MainThread:3226566 [wandb_init.py:init():813] calling init triggers +2025-09-14 02:51:13,489 INFO MainThread:3226566 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 02:51:13,489 INFO MainThread:3226566 [wandb_init.py:init():854] starting backend +2025-09-14 02:51:13,755 INFO MainThread:3226566 [wandb_init.py:init():857] sending inform_init request +2025-09-14 02:51:13,768 INFO MainThread:3226566 [wandb_init.py:init():865] backend started and connected +2025-09-14 02:51:13,790 INFO MainThread:3226566 [wandb_init.py:init():936] updated telemetry +2025-09-14 02:51:13,818 INFO MainThread:3226566 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 02:51:14,559 INFO MainThread:3226566 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 02:51:15,016 INFO MainThread:3226566 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 02:51:15,017 INFO MainThread:3226566 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 02:51:15,017 INFO MainThread:3226566 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 02:51:15,018 INFO MainThread:3226566 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 02:51:15,024 INFO MainThread:3226566 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 02:53:19,710 INFO wandb-AsyncioManager-main:3226566 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 02:53:19,712 INFO wandb-AsyncioManager-main:3226566 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_025113-gl43w77w/run-gl43w77w.wandb b/wandb/run-20250914_025113-gl43w77w/run-gl43w77w.wandb new file mode 100644 index 0000000000000000000000000000000000000000..12997afad9cb70b2e1c05226fa598ebf41752379 Binary files /dev/null and b/wandb/run-20250914_025113-gl43w77w/run-gl43w77w.wandb differ diff --git a/wandb/run-20250914_025738-ennija93/files/config.yaml b/wandb/run-20250914_025738-ennija93/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fd38700f80ed65fbbbe6e9fe098a7bfa3dd89310 --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/files/config.yaml @@ -0,0 +1,85 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + x4n09hv7clb3hlfo9pauktdqzfze6ba3: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=64 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.val_check_interval=10000 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533139656704" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T18:57:38.070037Z" + writerId: x4n09hv7clb3hlfo9pauktdqzfze6ba3 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_025738-ennija93/files/output.log b/wandb/run-20250914_025738-ennija93/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ad5d0c01703efb1a75332a42f060c7d484368d68 --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/files/output.log @@ -0,0 +1,91 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run + self.strategy.setup(self) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup + self.init_deepspeed() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed + self._initialize_deepspeed_train(self.model) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 492, in _initialize_deepspeed_train + ) = self._init_optimizers() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 468, in _init_optimizers + raise MisconfigurationException( +lightning.fabric.utilities.exceptions.MisconfigurationException: DeepSpeed currently only supports single optimizer, single optional scheduler. +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run +[rank0]: self.strategy.setup(self) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup +[rank0]: self.init_deepspeed() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed +[rank0]: self._initialize_deepspeed_train(self.model) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 492, in _initialize_deepspeed_train +[rank0]: ) = self._init_optimizers() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 468, in _init_optimizers +[rank0]: raise MisconfigurationException( +[rank0]: lightning.fabric.utilities.exceptions.MisconfigurationException: DeepSpeed currently only supports single optimizer, single optional scheduler. diff --git a/wandb/run-20250914_025738-ennija93/files/requirements.txt b/wandb/run-20250914_025738-ennija93/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..0f41d7d9bd6a8303fbd36805dd1e61ca08114074 --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/files/requirements.txt @@ -0,0 +1,137 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_025738-ennija93/files/wandb-metadata.json b/wandb/run-20250914_025738-ennija93/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e6087e5736996d5dd0cc656080c527b9204dca11 --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T18:57:38.070037Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533139656704" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "x4n09hv7clb3hlfo9pauktdqzfze6ba3" +} \ No newline at end of file diff --git a/wandb/run-20250914_025738-ennija93/files/wandb-summary.json b/wandb/run-20250914_025738-ennija93/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..9425ca838b51ff89a554781cd735aeae2cb8804f --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":46,"_wandb":{"runtime":46}} \ No newline at end of file diff --git a/wandb/run-20250914_025738-ennija93/logs/debug-core.log b/wandb/run-20250914_025738-ennija93/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..713937abb8741363672da837afbc051f3b12fdeb --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T02:57:38.172102919+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp0eg6l_w7/port-3250911.txt","pid":3250911,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T02:57:38.172357744+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T02:57:38.173135749+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3250911} +{"time":"2025-09-14T02:57:38.173125064+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3250911-3258510-1750902890/socket","Net":"unix"}} +{"time":"2025-09-14T02:57:38.333852438+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T02:57:38.351538499+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ennija93","id":"1(@)"} +{"time":"2025-09-14T02:57:38.854645614+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ennija93","id":"1(@)"} +{"time":"2025-09-14T02:58:25.546842283+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T02:58:25.547647797+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T02:58:25.547765095+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T02:58:25.547819682+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T02:58:25.548529875+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3250911-3258510-1750902890/socket","Net":"unix"}} +{"time":"2025-09-14T02:58:27.113214245+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T02:58:27.113317559+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T02:58:27.113349294+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_025738-ennija93/logs/debug-internal.log b/wandb/run-20250914_025738-ennija93/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..27425a11d43ea31895aa97954e9c7e1cf325d7c3 --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T02:57:38.351796156+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T02:57:38.39490858+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T02:57:38.85450886+08:00","level":"INFO","msg":"stream: created new stream","id":"ennija93"} +{"time":"2025-09-14T02:57:38.854629904+08:00","level":"INFO","msg":"stream: started","id":"ennija93"} +{"time":"2025-09-14T02:57:38.854645287+08:00","level":"INFO","msg":"writer: started","stream_id":"ennija93"} +{"time":"2025-09-14T02:57:38.854660727+08:00","level":"INFO","msg":"handler: started","stream_id":"ennija93"} +{"time":"2025-09-14T02:57:38.854687844+08:00","level":"INFO","msg":"sender: started","stream_id":"ennija93"} +{"time":"2025-09-14T02:58:25.547599099+08:00","level":"INFO","msg":"stream: closing","id":"ennija93"} +{"time":"2025-09-14T02:58:26.839276163+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T02:58:27.110615534+08:00","level":"INFO","msg":"handler: closed","stream_id":"ennija93"} +{"time":"2025-09-14T02:58:27.1115659+08:00","level":"INFO","msg":"sender: closed","stream_id":"ennija93"} +{"time":"2025-09-14T02:58:27.111652748+08:00","level":"INFO","msg":"stream: closed","id":"ennija93"} diff --git a/wandb/run-20250914_025738-ennija93/logs/debug.log b/wandb/run-20250914_025738-ennija93/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..80deec7749cf93b1107cf56677c98d64d96f67d4 --- /dev/null +++ b/wandb/run-20250914_025738-ennija93/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 02:57:38,078 INFO MainThread:3250911 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 02:57:38,078 INFO MainThread:3250911 [wandb_setup.py:_flush():81] Configure stats pid to 3250911 +2025-09-14 02:57:38,078 INFO MainThread:3250911 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 02:57:38,079 INFO MainThread:3250911 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 02:57:38,079 INFO MainThread:3250911 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 02:57:38,079 INFO MainThread:3250911 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_025738-ennija93/logs/debug.log +2025-09-14 02:57:38,080 INFO MainThread:3250911 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_025738-ennija93/logs/debug-internal.log +2025-09-14 02:57:38,080 INFO MainThread:3250911 [wandb_init.py:init():813] calling init triggers +2025-09-14 02:57:38,080 INFO MainThread:3250911 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 02:57:38,081 INFO MainThread:3250911 [wandb_init.py:init():854] starting backend +2025-09-14 02:57:38,334 INFO MainThread:3250911 [wandb_init.py:init():857] sending inform_init request +2025-09-14 02:57:38,346 INFO MainThread:3250911 [wandb_init.py:init():865] backend started and connected +2025-09-14 02:57:38,353 INFO MainThread:3250911 [wandb_init.py:init():936] updated telemetry +2025-09-14 02:57:38,387 INFO MainThread:3250911 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 02:57:39,209 INFO MainThread:3250911 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 02:57:39,651 INFO MainThread:3250911 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 02:57:39,652 INFO MainThread:3250911 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 02:57:39,652 INFO MainThread:3250911 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 02:57:39,652 INFO MainThread:3250911 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 02:57:39,659 INFO MainThread:3250911 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 02:58:25,545 INFO wandb-AsyncioManager-main:3250911 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 02:58:25,546 INFO wandb-AsyncioManager-main:3250911 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_025738-ennija93/run-ennija93.wandb b/wandb/run-20250914_025738-ennija93/run-ennija93.wandb new file mode 100644 index 0000000000000000000000000000000000000000..b9e5ebb7ec78aae1b3eb3a39ca923f8cd4d9c3a8 Binary files /dev/null and b/wandb/run-20250914_025738-ennija93/run-ennija93.wandb differ diff --git a/wandb/run-20250914_030446-ytsr6edl/files/output.log b/wandb/run-20250914_030446-ytsr6edl/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b9a5deff90331b1705af0f0da0ddba4ac861099c --- /dev/null +++ b/wandb/run-20250914_030446-ytsr6edl/files/output.log @@ -0,0 +1,4 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +[rank: 6] Child process with PID 3275115 terminated with code 1. Forcefully terminating all other processes to avoid zombies 🧟 diff --git a/wandb/run-20250914_030446-ytsr6edl/files/requirements.txt b/wandb/run-20250914_030446-ytsr6edl/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..04d92613668eed2012d0762223a6685fb471c959 --- /dev/null +++ b/wandb/run-20250914_030446-ytsr6edl/files/requirements.txt @@ -0,0 +1,121 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 diff --git a/wandb/run-20250914_030446-ytsr6edl/files/wandb-metadata.json b/wandb/run-20250914_030446-ytsr6edl/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e5e796b265a05e4d37b2520a67fe7b063b4d7514 --- /dev/null +++ b/wandb/run-20250914_030446-ytsr6edl/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T19:04:46.600338Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=64", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=ddp_find_unused_parameters_true", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533139861504" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "1wc7bsws09n3x9nt1y6ry0hrg6h682mv" +} \ No newline at end of file diff --git a/wandb/run-20250914_030446-ytsr6edl/logs/debug-core.log b/wandb/run-20250914_030446-ytsr6edl/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..e772b8be08a46be5cea26a6154acc622e579fec1 --- /dev/null +++ b/wandb/run-20250914_030446-ytsr6edl/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T03:04:46.701491619+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpcro21f52/port-3273942.txt","pid":3273942,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T03:04:46.70208328+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T03:04:46.702742946+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3273942} +{"time":"2025-09-14T03:04:46.702730439+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3273942-3278227-4240016479/socket","Net":"unix"}} +{"time":"2025-09-14T03:04:46.864378747+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T03:04:46.89086533+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ytsr6edl","id":"1(@)"} +{"time":"2025-09-14T03:04:47.411034489+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ytsr6edl","id":"1(@)"} +{"time":"2025-09-14T03:06:02.531765476+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_030446-ytsr6edl/logs/debug-internal.log b/wandb/run-20250914_030446-ytsr6edl/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..006686998389dc01652ab1131b68f8b032855049 --- /dev/null +++ b/wandb/run-20250914_030446-ytsr6edl/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T03:04:46.891262833+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T03:04:46.93862831+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T03:04:47.410910715+08:00","level":"INFO","msg":"stream: created new stream","id":"ytsr6edl"} +{"time":"2025-09-14T03:04:47.411020419+08:00","level":"INFO","msg":"stream: started","id":"ytsr6edl"} +{"time":"2025-09-14T03:04:47.411051249+08:00","level":"INFO","msg":"writer: started","stream_id":"ytsr6edl"} +{"time":"2025-09-14T03:04:47.411064766+08:00","level":"INFO","msg":"handler: started","stream_id":"ytsr6edl"} +{"time":"2025-09-14T03:04:47.411112685+08:00","level":"INFO","msg":"sender: started","stream_id":"ytsr6edl"} diff --git a/wandb/run-20250914_030446-ytsr6edl/logs/debug.log b/wandb/run-20250914_030446-ytsr6edl/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..75f602b4c88dba948d2b793bb931e96462686059 --- /dev/null +++ b/wandb/run-20250914_030446-ytsr6edl/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 03:04:46,606 INFO MainThread:3273942 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 03:04:46,607 INFO MainThread:3273942 [wandb_setup.py:_flush():81] Configure stats pid to 3273942 +2025-09-14 03:04:46,607 INFO MainThread:3273942 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 03:04:46,607 INFO MainThread:3273942 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 03:04:46,608 INFO MainThread:3273942 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 03:04:46,608 INFO MainThread:3273942 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_030446-ytsr6edl/logs/debug.log +2025-09-14 03:04:46,609 INFO MainThread:3273942 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_030446-ytsr6edl/logs/debug-internal.log +2025-09-14 03:04:46,609 INFO MainThread:3273942 [wandb_init.py:init():813] calling init triggers +2025-09-14 03:04:46,609 INFO MainThread:3273942 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 03:04:46,610 INFO MainThread:3273942 [wandb_init.py:init():854] starting backend +2025-09-14 03:04:46,866 INFO MainThread:3273942 [wandb_init.py:init():857] sending inform_init request +2025-09-14 03:04:46,876 INFO MainThread:3273942 [wandb_init.py:init():865] backend started and connected +2025-09-14 03:04:46,886 INFO MainThread:3273942 [wandb_init.py:init():936] updated telemetry +2025-09-14 03:04:46,932 INFO MainThread:3273942 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 03:04:47,757 INFO MainThread:3273942 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 03:04:48,109 INFO MainThread:3273942 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 03:04:48,109 INFO MainThread:3273942 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 03:04:48,109 INFO MainThread:3273942 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 03:04:48,109 INFO MainThread:3273942 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 03:04:48,116 INFO MainThread:3273942 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_030446-ytsr6edl/run-ytsr6edl.wandb b/wandb/run-20250914_030446-ytsr6edl/run-ytsr6edl.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_030720-i9qj4du3/files/output.log b/wandb/run-20250914_030720-i9qj4du3/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..d8d33574ce5396e9b7424ec278c849d36ed5aa5d --- /dev/null +++ b/wandb/run-20250914_030720-i9qj4du3/files/output.log @@ -0,0 +1,3 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] diff --git a/wandb/run-20250914_030720-i9qj4du3/files/requirements.txt b/wandb/run-20250914_030720-i9qj4du3/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..04d92613668eed2012d0762223a6685fb471c959 --- /dev/null +++ b/wandb/run-20250914_030720-i9qj4du3/files/requirements.txt @@ -0,0 +1,121 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 diff --git a/wandb/run-20250914_030720-i9qj4du3/files/wandb-metadata.json b/wandb/run-20250914_030720-i9qj4du3/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..1ce436c509ba3f038ab5156750bfc44c6c1f5015 --- /dev/null +++ b/wandb/run-20250914_030720-i9qj4du3/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T19:07:20.845754Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=1", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=ddp_find_unused_parameters_true", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533140205568" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "65n657fmfr840en3kmgcb3jqdztfplrl" +} \ No newline at end of file diff --git a/wandb/run-20250914_030720-i9qj4du3/logs/debug-core.log b/wandb/run-20250914_030720-i9qj4du3/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..2de1af5d8a432e8a8b70592a98e55dc265403b5c --- /dev/null +++ b/wandb/run-20250914_030720-i9qj4du3/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T03:07:20.913389092+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpugfa7gnh/port-3283392.txt","pid":3283392,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T03:07:20.913575375+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T03:07:20.914294664+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3283392} +{"time":"2025-09-14T03:07:20.914289202+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3283392-3287108-3182837382/socket","Net":"unix"}} +{"time":"2025-09-14T03:07:21.103914786+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T03:07:21.141447686+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"i9qj4du3","id":"1(@)"} +{"time":"2025-09-14T03:07:21.627768179+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"i9qj4du3","id":"1(@)"} +{"time":"2025-09-14T03:09:03.715774262+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_030720-i9qj4du3/logs/debug-internal.log b/wandb/run-20250914_030720-i9qj4du3/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..fc648d96b07d759e960f267a7227513708e60295 --- /dev/null +++ b/wandb/run-20250914_030720-i9qj4du3/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T03:07:21.141590847+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T03:07:21.171010972+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T03:07:21.627591632+08:00","level":"INFO","msg":"stream: created new stream","id":"i9qj4du3"} +{"time":"2025-09-14T03:07:21.627752322+08:00","level":"INFO","msg":"stream: started","id":"i9qj4du3"} +{"time":"2025-09-14T03:07:21.627752276+08:00","level":"INFO","msg":"writer: started","stream_id":"i9qj4du3"} +{"time":"2025-09-14T03:07:21.627766073+08:00","level":"INFO","msg":"handler: started","stream_id":"i9qj4du3"} +{"time":"2025-09-14T03:07:21.627868374+08:00","level":"INFO","msg":"sender: started","stream_id":"i9qj4du3"} diff --git a/wandb/run-20250914_030720-i9qj4du3/logs/debug.log b/wandb/run-20250914_030720-i9qj4du3/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..48567825c52b0cbb388a922184fad7c1b635b950 --- /dev/null +++ b/wandb/run-20250914_030720-i9qj4du3/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 03:07:20,851 INFO MainThread:3283392 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 03:07:20,852 INFO MainThread:3283392 [wandb_setup.py:_flush():81] Configure stats pid to 3283392 +2025-09-14 03:07:20,852 INFO MainThread:3283392 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 03:07:20,852 INFO MainThread:3283392 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 03:07:20,852 INFO MainThread:3283392 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 03:07:20,853 INFO MainThread:3283392 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_030720-i9qj4du3/logs/debug.log +2025-09-14 03:07:20,853 INFO MainThread:3283392 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_030720-i9qj4du3/logs/debug-internal.log +2025-09-14 03:07:20,854 INFO MainThread:3283392 [wandb_init.py:init():813] calling init triggers +2025-09-14 03:07:20,854 INFO MainThread:3283392 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 03:07:20,854 INFO MainThread:3283392 [wandb_init.py:init():854] starting backend +2025-09-14 03:07:21,106 INFO MainThread:3283392 [wandb_init.py:init():857] sending inform_init request +2025-09-14 03:07:21,119 INFO MainThread:3283392 [wandb_init.py:init():865] backend started and connected +2025-09-14 03:07:21,140 INFO MainThread:3283392 [wandb_init.py:init():936] updated telemetry +2025-09-14 03:07:21,198 INFO MainThread:3283392 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 03:07:21,948 INFO MainThread:3283392 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 03:07:22,303 INFO MainThread:3283392 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 03:07:22,304 INFO MainThread:3283392 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 03:07:22,304 INFO MainThread:3283392 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 03:07:22,304 INFO MainThread:3283392 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 03:07:22,311 INFO MainThread:3283392 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_030720-i9qj4du3/run-i9qj4du3.wandb b/wandb/run-20250914_030720-i9qj4du3/run-i9qj4du3.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_031025-42udz6yn/files/output.log b/wandb/run-20250914_031025-42udz6yn/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..d8d33574ce5396e9b7424ec278c849d36ed5aa5d --- /dev/null +++ b/wandb/run-20250914_031025-42udz6yn/files/output.log @@ -0,0 +1,3 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] diff --git a/wandb/run-20250914_031025-42udz6yn/files/requirements.txt b/wandb/run-20250914_031025-42udz6yn/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..04d92613668eed2012d0762223a6685fb471c959 --- /dev/null +++ b/wandb/run-20250914_031025-42udz6yn/files/requirements.txt @@ -0,0 +1,121 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 diff --git a/wandb/run-20250914_031025-42udz6yn/files/wandb-metadata.json b/wandb/run-20250914_031025-42udz6yn/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bb8823b5d6938d4fb6e938f76e094b77289c2756 --- /dev/null +++ b/wandb/run-20250914_031025-42udz6yn/files/wandb-metadata.json @@ -0,0 +1,54 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T19:10:25.760416Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=1", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=fsdp", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533140799488" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ua2togbyp15o468pgowijvaw3n81pv2p" +} \ No newline at end of file diff --git a/wandb/run-20250914_031025-42udz6yn/logs/debug-core.log b/wandb/run-20250914_031025-42udz6yn/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..2a367c196950c3e523f0fb4721fdab7d1be23262 --- /dev/null +++ b/wandb/run-20250914_031025-42udz6yn/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T03:10:25.827296042+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpiyjxuzar/port-3291297.txt","pid":3291297,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T03:10:25.827484927+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T03:10:25.828209466+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3291297} +{"time":"2025-09-14T03:10:25.828204095+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3291297-3295382-167794407/socket","Net":"unix"}} +{"time":"2025-09-14T03:10:26.017675552+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T03:10:26.033380872+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"42udz6yn","id":"1(@)"} +{"time":"2025-09-14T03:10:26.537963052+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"42udz6yn","id":"1(@)"} +{"time":"2025-09-14T03:11:17.015440367+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_031025-42udz6yn/logs/debug-internal.log b/wandb/run-20250914_031025-42udz6yn/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0fe2396534582b41b604a584111c60ec562f4020 --- /dev/null +++ b/wandb/run-20250914_031025-42udz6yn/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T03:10:26.033656202+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T03:10:26.080311123+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T03:10:26.537836173+08:00","level":"INFO","msg":"stream: created new stream","id":"42udz6yn"} +{"time":"2025-09-14T03:10:26.537947846+08:00","level":"INFO","msg":"stream: started","id":"42udz6yn"} +{"time":"2025-09-14T03:10:26.537976477+08:00","level":"INFO","msg":"writer: started","stream_id":"42udz6yn"} +{"time":"2025-09-14T03:10:26.537991194+08:00","level":"INFO","msg":"sender: started","stream_id":"42udz6yn"} +{"time":"2025-09-14T03:10:26.538110171+08:00","level":"INFO","msg":"handler: started","stream_id":"42udz6yn"} diff --git a/wandb/run-20250914_031025-42udz6yn/logs/debug.log b/wandb/run-20250914_031025-42udz6yn/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..bfee9836a7bfd1a0cb12609b29b0f6454753fc67 --- /dev/null +++ b/wandb/run-20250914_031025-42udz6yn/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 03:10:25,767 INFO MainThread:3291297 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 03:10:25,767 INFO MainThread:3291297 [wandb_setup.py:_flush():81] Configure stats pid to 3291297 +2025-09-14 03:10:25,768 INFO MainThread:3291297 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 03:10:25,768 INFO MainThread:3291297 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 03:10:25,768 INFO MainThread:3291297 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 03:10:25,769 INFO MainThread:3291297 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_031025-42udz6yn/logs/debug.log +2025-09-14 03:10:25,769 INFO MainThread:3291297 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_031025-42udz6yn/logs/debug-internal.log +2025-09-14 03:10:25,769 INFO MainThread:3291297 [wandb_init.py:init():813] calling init triggers +2025-09-14 03:10:25,770 INFO MainThread:3291297 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 03:10:25,770 INFO MainThread:3291297 [wandb_init.py:init():854] starting backend +2025-09-14 03:10:26,018 INFO MainThread:3291297 [wandb_init.py:init():857] sending inform_init request +2025-09-14 03:10:26,028 INFO MainThread:3291297 [wandb_init.py:init():865] backend started and connected +2025-09-14 03:10:26,034 INFO MainThread:3291297 [wandb_init.py:init():936] updated telemetry +2025-09-14 03:10:26,068 INFO MainThread:3291297 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 03:10:26,838 INFO MainThread:3291297 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 03:10:27,183 INFO MainThread:3291297 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 03:10:27,184 INFO MainThread:3291297 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 03:10:27,184 INFO MainThread:3291297 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 03:10:27,184 INFO MainThread:3291297 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 03:10:27,191 INFO MainThread:3291297 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_031025-42udz6yn/run-42udz6yn.wandb b/wandb/run-20250914_031025-42udz6yn/run-42udz6yn.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_031430-8xc03wmj/files/config.yaml b/wandb/run-20250914_031430-8xc03wmj/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fb23c5fccacfed5bcdfb9d80ca3acef37527ed30 --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/files/config.yaml @@ -0,0 +1,86 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + 2lxujrz9n31p8vxb4nzlh4irqt9sq8ip: + args: + - fit + - --data=RSAGame + - --data.batch_size=1 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.val_check_interval=10000 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533141336064" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T19:14:30.794287Z" + writerId: 2lxujrz9n31p8vxb4nzlh4irqt9sq8ip + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_031430-8xc03wmj/files/output.log b/wandb/run-20250914_031430-8xc03wmj/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..a2470e096de0daa3561009e5c5a661fda4bc896d --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/files/output.log @@ -0,0 +1,93 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +[DEBUG] Optimizer created with 2 parameter groups + Group 0: LR=1e-05, Params=443 + Group 1: LR=1e-05, Params=446 +Parameter Offload - Persistent parameters statistics: param_count = 441, numel = 689160 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/utilities/model_summary/model_summary.py:231: Precision bf16-mixed is not supported by the model summary. Estimated model size in MB will not be accurate. Using 32 bits instead. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 991, in _run + call._call_callback_hooks(self, "on_fit_start") + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks + fn(trainer, trainer.lightning_module, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_summary.py", line 65, in on_fit_start + summary_data = model_summary._get_summary_data() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/utilities/model_summary/model_summary_deepspeed.py", line 102, in _get_summary_data + ("FLOPs", list(map(get_human_readable_count, (sum(x.values()) for x in self.flop_counts.values())))), +AttributeError: 'DeepSpeedSummary' object has no attribute 'flop_counts' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 991, in _run +[rank0]: call._call_callback_hooks(self, "on_fit_start") +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks +[rank0]: fn(trainer, trainer.lightning_module, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_summary.py", line 65, in on_fit_start +[rank0]: summary_data = model_summary._get_summary_data() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/utilities/model_summary/model_summary_deepspeed.py", line 102, in _get_summary_data +[rank0]: ("FLOPs", list(map(get_human_readable_count, (sum(x.values()) for x in self.flop_counts.values())))), +[rank0]: AttributeError: 'DeepSpeedSummary' object has no attribute 'flop_counts' diff --git a/wandb/run-20250914_031430-8xc03wmj/files/requirements.txt b/wandb/run-20250914_031430-8xc03wmj/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..0f41d7d9bd6a8303fbd36805dd1e61ca08114074 --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/files/requirements.txt @@ -0,0 +1,137 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_031430-8xc03wmj/files/wandb-metadata.json b/wandb/run-20250914_031430-8xc03wmj/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6c2d1a96ccd13b9aa0039ea22dd31391f5a55fd5 --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/files/wandb-metadata.json @@ -0,0 +1,55 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T19:14:30.794287Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=1", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533141336064" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "2lxujrz9n31p8vxb4nzlh4irqt9sq8ip" +} \ No newline at end of file diff --git a/wandb/run-20250914_031430-8xc03wmj/files/wandb-summary.json b/wandb/run-20250914_031430-8xc03wmj/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..022646b35b7c26b16cc2e065b40d7c01e5cfafde --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":76},"_runtime":76} \ No newline at end of file diff --git a/wandb/run-20250914_031430-8xc03wmj/logs/debug-core.log b/wandb/run-20250914_031430-8xc03wmj/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..c9d8c5771e2d8d05b0fbfa94f5e17dda952671af --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T03:14:30.896926258+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp6o4g6wb0/port-3303777.txt","pid":3303777,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T03:14:30.897830491+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T03:14:30.898567534+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3303777} +{"time":"2025-09-14T03:14:30.898763179+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3303777-3330371-1878067033/socket","Net":"unix"}} +{"time":"2025-09-14T03:14:31.049021277+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T03:14:31.067832122+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"8xc03wmj","id":"1(@)"} +{"time":"2025-09-14T03:14:31.570930522+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"8xc03wmj","id":"1(@)"} +{"time":"2025-09-14T03:15:48.274438972+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T03:15:48.274716557+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T03:15:48.274771101+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T03:15:48.274871673+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T03:15:48.275439891+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3303777-3330371-1878067033/socket","Net":"unix"}} +{"time":"2025-09-14T03:15:49.814257711+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T03:15:49.814298677+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T03:15:49.814316428+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_031430-8xc03wmj/logs/debug-internal.log b/wandb/run-20250914_031430-8xc03wmj/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..718426bcc037fdb5dfd999ef15e577292624ead8 --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T03:14:31.068088112+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T03:14:31.11284527+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T03:14:31.570818114+08:00","level":"INFO","msg":"stream: created new stream","id":"8xc03wmj"} +{"time":"2025-09-14T03:14:31.570915994+08:00","level":"INFO","msg":"stream: started","id":"8xc03wmj"} +{"time":"2025-09-14T03:14:31.570963702+08:00","level":"INFO","msg":"handler: started","stream_id":"8xc03wmj"} +{"time":"2025-09-14T03:14:31.570949265+08:00","level":"INFO","msg":"writer: started","stream_id":"8xc03wmj"} +{"time":"2025-09-14T03:14:31.571006257+08:00","level":"INFO","msg":"sender: started","stream_id":"8xc03wmj"} +{"time":"2025-09-14T03:15:48.274618404+08:00","level":"INFO","msg":"stream: closing","id":"8xc03wmj"} +{"time":"2025-09-14T03:15:49.555827419+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T03:15:49.812957217+08:00","level":"INFO","msg":"handler: closed","stream_id":"8xc03wmj"} +{"time":"2025-09-14T03:15:49.813630891+08:00","level":"INFO","msg":"sender: closed","stream_id":"8xc03wmj"} +{"time":"2025-09-14T03:15:49.813691403+08:00","level":"INFO","msg":"stream: closed","id":"8xc03wmj"} diff --git a/wandb/run-20250914_031430-8xc03wmj/logs/debug.log b/wandb/run-20250914_031430-8xc03wmj/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..65e32fc6a4a03f21f7f32d5e2e12cf48394bb2bb --- /dev/null +++ b/wandb/run-20250914_031430-8xc03wmj/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 03:14:30,801 INFO MainThread:3303777 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 03:14:30,801 INFO MainThread:3303777 [wandb_setup.py:_flush():81] Configure stats pid to 3303777 +2025-09-14 03:14:30,801 INFO MainThread:3303777 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 03:14:30,801 INFO MainThread:3303777 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 03:14:30,802 INFO MainThread:3303777 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 03:14:30,802 INFO MainThread:3303777 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_031430-8xc03wmj/logs/debug.log +2025-09-14 03:14:30,802 INFO MainThread:3303777 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_031430-8xc03wmj/logs/debug-internal.log +2025-09-14 03:14:30,803 INFO MainThread:3303777 [wandb_init.py:init():813] calling init triggers +2025-09-14 03:14:30,803 INFO MainThread:3303777 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 03:14:30,803 INFO MainThread:3303777 [wandb_init.py:init():854] starting backend +2025-09-14 03:14:31,050 INFO MainThread:3303777 [wandb_init.py:init():857] sending inform_init request +2025-09-14 03:14:31,061 INFO MainThread:3303777 [wandb_init.py:init():865] backend started and connected +2025-09-14 03:14:31,068 INFO MainThread:3303777 [wandb_init.py:init():936] updated telemetry +2025-09-14 03:14:31,094 INFO MainThread:3303777 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 03:14:31,845 INFO MainThread:3303777 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 03:14:32,236 INFO MainThread:3303777 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 03:14:32,237 INFO MainThread:3303777 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 03:14:32,237 INFO MainThread:3303777 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 03:14:32,237 INFO MainThread:3303777 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 03:14:32,243 INFO MainThread:3303777 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 03:15:48,274 INFO wandb-AsyncioManager-main:3303777 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 03:15:48,275 INFO wandb-AsyncioManager-main:3303777 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_031430-8xc03wmj/run-8xc03wmj.wandb b/wandb/run-20250914_031430-8xc03wmj/run-8xc03wmj.wandb new file mode 100644 index 0000000000000000000000000000000000000000..9bbacf82a700f707f92f5f460f82c35667009781 Binary files /dev/null and b/wandb/run-20250914_031430-8xc03wmj/run-8xc03wmj.wandb differ diff --git a/wandb/run-20250914_032121-llcw3uj7/files/output.log b/wandb/run-20250914_032121-llcw3uj7/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..fc3e97a86bf6da3d5d5a0ee6be5566ff60e18484 --- /dev/null +++ b/wandb/run-20250914_032121-llcw3uj7/files/output.log @@ -0,0 +1,94 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +[DEBUG] Optimizer created with 2 parameter groups + Group 0: LR=1e-05, Params=443 + Group 1: LR=1e-05, Params=446 +Parameter Offload - Persistent parameters statistics: param_count = 441, numel = 689160 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/utilities/model_summary/model_summary.py:231: Precision bf16-mixed is not supported by the model summary. Estimated model size in MB will not be accurate. Using 32 bits instead. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 991, in _run + call._call_callback_hooks(self, "on_fit_start") + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks + fn(trainer, trainer.lightning_module, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_summary.py", line 65, in on_fit_start + summary_data = model_summary._get_summary_data() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/utilities/model_summary/model_summary_deepspeed.py", line 102, in _get_summary_data + ("FLOPs", list(map(get_human_readable_count, (sum(x.values()) for x in self.flop_counts.values())))), +AttributeError: 'DeepSpeedSummary' object has no attribute 'flop_counts' +[rank: 1] Child process with PID 3344110 terminated with code 1. Forcefully terminating all other processes to avoid zombies 🧟 +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 991, in _run +[rank0]: call._call_callback_hooks(self, "on_fit_start") +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks +[rank0]: fn(trainer, trainer.lightning_module, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_summary.py", line 65, in on_fit_start +[rank0]: summary_data = model_summary._get_summary_data() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/utilities/model_summary/model_summary_deepspeed.py", line 102, in _get_summary_data +[rank0]: ("FLOPs", list(map(get_human_readable_count, (sum(x.values()) for x in self.flop_counts.values())))), +[rank0]: AttributeError: 'DeepSpeedSummary' object has no attribute 'flop_counts' diff --git a/wandb/run-20250914_032121-llcw3uj7/files/requirements.txt b/wandb/run-20250914_032121-llcw3uj7/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..0f41d7d9bd6a8303fbd36805dd1e61ca08114074 --- /dev/null +++ b/wandb/run-20250914_032121-llcw3uj7/files/requirements.txt @@ -0,0 +1,137 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_032121-llcw3uj7/files/wandb-metadata.json b/wandb/run-20250914_032121-llcw3uj7/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8951c97ad057e851ce152e7ea9d294ce04ef612e --- /dev/null +++ b/wandb/run-20250914_032121-llcw3uj7/files/wandb-metadata.json @@ -0,0 +1,55 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T19:21:21.519133Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=1", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=10000" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533120360448" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "7joghvq5y3m91xiw4pc8re1f6z2ou6oh" +} \ No newline at end of file diff --git a/wandb/run-20250914_032121-llcw3uj7/logs/debug-core.log b/wandb/run-20250914_032121-llcw3uj7/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bce91c78fa849859461c214fd2411e1a17bdb4d2 --- /dev/null +++ b/wandb/run-20250914_032121-llcw3uj7/logs/debug-core.log @@ -0,0 +1,9 @@ +{"time":"2025-09-14T03:21:21.563647062+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpaw1am2hy/port-3343073.txt","pid":3343073,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T03:21:21.563852411+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T03:21:21.564621485+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3343073} +{"time":"2025-09-14T03:21:21.564634035+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3343073-3349891-1928923241/socket","Net":"unix"}} +{"time":"2025-09-14T03:21:21.752701381+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T03:21:21.764007617+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"llcw3uj7","id":"1(@)"} +{"time":"2025-09-14T03:21:22.252729216+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"llcw3uj7","id":"1(@)"} +{"time":"2025-09-14T03:22:54.25700781+08:00","level":"ERROR","msg":"connection: fatal error reading connection","error":"read unix /tmp/wandb-3343073-3349891-1928923241/socket->@: read: connection reset by peer","id":"1(@)"} +{"time":"2025-09-14T03:22:55.046380446+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_032121-llcw3uj7/logs/debug-internal.log b/wandb/run-20250914_032121-llcw3uj7/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..05ab3ec100710638c1161e466fb8359222aaa053 --- /dev/null +++ b/wandb/run-20250914_032121-llcw3uj7/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T03:21:21.764249762+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T03:21:21.801468876+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T03:21:22.252643219+08:00","level":"INFO","msg":"stream: created new stream","id":"llcw3uj7"} +{"time":"2025-09-14T03:21:22.252714755+08:00","level":"INFO","msg":"stream: started","id":"llcw3uj7"} +{"time":"2025-09-14T03:21:22.252773703+08:00","level":"INFO","msg":"writer: started","stream_id":"llcw3uj7"} +{"time":"2025-09-14T03:21:22.2527884+08:00","level":"INFO","msg":"sender: started","stream_id":"llcw3uj7"} +{"time":"2025-09-14T03:21:22.252829164+08:00","level":"INFO","msg":"handler: started","stream_id":"llcw3uj7"} diff --git a/wandb/run-20250914_032121-llcw3uj7/logs/debug.log b/wandb/run-20250914_032121-llcw3uj7/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a9ce556e4131cb6af9e172b61a0d7c2f5440a213 --- /dev/null +++ b/wandb/run-20250914_032121-llcw3uj7/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 03:21:21,523 INFO MainThread:3343073 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 03:21:21,523 INFO MainThread:3343073 [wandb_setup.py:_flush():81] Configure stats pid to 3343073 +2025-09-14 03:21:21,523 INFO MainThread:3343073 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 03:21:21,523 INFO MainThread:3343073 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 03:21:21,524 INFO MainThread:3343073 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 03:21:21,524 INFO MainThread:3343073 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_032121-llcw3uj7/logs/debug.log +2025-09-14 03:21:21,524 INFO MainThread:3343073 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_032121-llcw3uj7/logs/debug-internal.log +2025-09-14 03:21:21,524 INFO MainThread:3343073 [wandb_init.py:init():813] calling init triggers +2025-09-14 03:21:21,524 INFO MainThread:3343073 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 03:21:21,524 INFO MainThread:3343073 [wandb_init.py:init():854] starting backend +2025-09-14 03:21:21,754 INFO MainThread:3343073 [wandb_init.py:init():857] sending inform_init request +2025-09-14 03:21:21,759 INFO MainThread:3343073 [wandb_init.py:init():865] backend started and connected +2025-09-14 03:21:21,764 INFO MainThread:3343073 [wandb_init.py:init():936] updated telemetry +2025-09-14 03:21:21,778 INFO MainThread:3343073 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 03:21:22,530 INFO MainThread:3343073 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 03:21:22,945 INFO MainThread:3343073 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 03:21:22,945 INFO MainThread:3343073 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 03:21:22,946 INFO MainThread:3343073 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 03:21:22,946 INFO MainThread:3343073 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 03:21:22,951 INFO MainThread:3343073 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_032121-llcw3uj7/run-llcw3uj7.wandb b/wandb/run-20250914_032121-llcw3uj7/run-llcw3uj7.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_032551-lcnpio7h/files/output.log b/wandb/run-20250914_032551-lcnpio7h/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7ddf37bb59a2087ac6793b8e97023f78e2b784fb --- /dev/null +++ b/wandb/run-20250914_032551-lcnpio7h/files/output.log @@ -0,0 +1,10 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +[DEBUG] Optimizer created with 2 parameter groups + Group 0: LR=1e-05, Params=443 + Group 1: LR=1e-05, Params=446 +Parameter Offload - Persistent parameters statistics: param_count = 441, numel = 689160 +Sanity Checking DataLoader 0: 0%| | 0/1 [00:00 + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 208, in run + self.setup_data() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 289, in setup_data + raise ValueError( +ValueError: `val_check_interval` (90000) must be less than or equal to the number of the training batches (2). If you want to disable validation set `limit_val_batches` to 0.0 instead. If you want to validate based on the total training batches, set `check_val_every_n_epoch=None`. +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 208, in run +[rank0]: self.setup_data() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 289, in setup_data +[rank0]: raise ValueError( +[rank0]: ValueError: `val_check_interval` (90000) must be less than or equal to the number of the training batches (2). If you want to disable validation set `limit_val_batches` to 0.0 instead. If you want to validate based on the total training batches, set `check_val_every_n_epoch=None`. diff --git a/wandb/run-20250914_044913-7y4j7voo/files/requirements.txt b/wandb/run-20250914_044913-7y4j7voo/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_044913-7y4j7voo/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_044913-7y4j7voo/files/wandb-metadata.json b/wandb/run-20250914_044913-7y4j7voo/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4de4610445ddcccd19654b8755badf1cff18b081 --- /dev/null +++ b/wandb/run-20250914_044913-7y4j7voo/files/wandb-metadata.json @@ -0,0 +1,57 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T20:49:13.747060Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=90000", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533430734848" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "dqi9e2n10xpabrlei1wrjbevdl1nzksb" +} \ No newline at end of file diff --git a/wandb/run-20250914_044913-7y4j7voo/files/wandb-summary.json b/wandb/run-20250914_044913-7y4j7voo/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..69e8a87828be38513e2aff6e058e4a9485d68a00 --- /dev/null +++ b/wandb/run-20250914_044913-7y4j7voo/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":38},"_runtime":38} \ No newline at end of file diff --git a/wandb/run-20250914_044913-7y4j7voo/logs/debug-core.log b/wandb/run-20250914_044913-7y4j7voo/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..1803bf1dc59230ae2ff837e4a368da121fb952da --- /dev/null +++ b/wandb/run-20250914_044913-7y4j7voo/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T04:49:13.81797964+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpsdvv9lqf/port-3693683.txt","pid":3693683,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T04:49:13.818266225+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T04:49:13.819363973+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3693683} +{"time":"2025-09-14T04:49:13.819321736+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3693683-3703572-3588825090/socket","Net":"unix"}} +{"time":"2025-09-14T04:49:14.009410354+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T04:49:14.02678386+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"7y4j7voo","id":"1(@)"} +{"time":"2025-09-14T04:49:14.549257453+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"7y4j7voo","id":"1(@)"} +{"time":"2025-09-14T04:49:52.976939295+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T04:49:52.977230174+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T04:49:52.977226557+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T04:49:52.977401495+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T04:49:52.97766203+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3693683-3703572-3588825090/socket","Net":"unix"}} +{"time":"2025-09-14T04:49:54.515392656+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T04:49:54.515433124+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T04:49:54.515442625+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_044913-7y4j7voo/logs/debug-internal.log b/wandb/run-20250914_044913-7y4j7voo/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..92b802858fe7489adfc7ea7361cd7d52314e0237 --- /dev/null +++ b/wandb/run-20250914_044913-7y4j7voo/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T04:49:14.027046149+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T04:49:14.072851934+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T04:49:14.549164787+08:00","level":"INFO","msg":"stream: created new stream","id":"7y4j7voo"} +{"time":"2025-09-14T04:49:14.549242687+08:00","level":"INFO","msg":"stream: started","id":"7y4j7voo"} +{"time":"2025-09-14T04:49:14.549271115+08:00","level":"INFO","msg":"writer: started","stream_id":"7y4j7voo"} +{"time":"2025-09-14T04:49:14.549290493+08:00","level":"INFO","msg":"handler: started","stream_id":"7y4j7voo"} +{"time":"2025-09-14T04:49:14.549318328+08:00","level":"INFO","msg":"sender: started","stream_id":"7y4j7voo"} +{"time":"2025-09-14T04:49:52.97706033+08:00","level":"INFO","msg":"stream: closing","id":"7y4j7voo"} +{"time":"2025-09-14T04:49:54.254000668+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T04:49:54.514416018+08:00","level":"INFO","msg":"handler: closed","stream_id":"7y4j7voo"} +{"time":"2025-09-14T04:49:54.514861027+08:00","level":"INFO","msg":"sender: closed","stream_id":"7y4j7voo"} +{"time":"2025-09-14T04:49:54.514897052+08:00","level":"INFO","msg":"stream: closed","id":"7y4j7voo"} diff --git a/wandb/run-20250914_044913-7y4j7voo/logs/debug.log b/wandb/run-20250914_044913-7y4j7voo/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..8f05ca803ccd043983c7922eb5f12343475325e5 --- /dev/null +++ b/wandb/run-20250914_044913-7y4j7voo/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 04:49:13,755 INFO MainThread:3693683 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 04:49:13,756 INFO MainThread:3693683 [wandb_setup.py:_flush():81] Configure stats pid to 3693683 +2025-09-14 04:49:13,756 INFO MainThread:3693683 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 04:49:13,757 INFO MainThread:3693683 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 04:49:13,757 INFO MainThread:3693683 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 04:49:13,757 INFO MainThread:3693683 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_044913-7y4j7voo/logs/debug.log +2025-09-14 04:49:13,758 INFO MainThread:3693683 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_044913-7y4j7voo/logs/debug-internal.log +2025-09-14 04:49:13,758 INFO MainThread:3693683 [wandb_init.py:init():813] calling init triggers +2025-09-14 04:49:13,758 INFO MainThread:3693683 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 04:49:13,759 INFO MainThread:3693683 [wandb_init.py:init():854] starting backend +2025-09-14 04:49:14,010 INFO MainThread:3693683 [wandb_init.py:init():857] sending inform_init request +2025-09-14 04:49:14,022 INFO MainThread:3693683 [wandb_init.py:init():865] backend started and connected +2025-09-14 04:49:14,032 INFO MainThread:3693683 [wandb_init.py:init():936] updated telemetry +2025-09-14 04:49:14,058 INFO MainThread:3693683 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 04:49:14,842 INFO MainThread:3693683 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 04:49:15,235 INFO MainThread:3693683 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 04:49:15,235 INFO MainThread:3693683 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 04:49:15,235 INFO MainThread:3693683 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 04:49:15,236 INFO MainThread:3693683 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 04:49:15,240 INFO MainThread:3693683 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 04:49:46,780 INFO MainThread:3693683 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 04:49:52,977 INFO wandb-AsyncioManager-main:3693683 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 04:49:52,977 INFO wandb-AsyncioManager-main:3693683 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_044913-7y4j7voo/run-7y4j7voo.wandb b/wandb/run-20250914_044913-7y4j7voo/run-7y4j7voo.wandb new file mode 100644 index 0000000000000000000000000000000000000000..47b2b3ebb8bda31059ede5c3135514bebd8e6338 Binary files /dev/null and b/wandb/run-20250914_044913-7y4j7voo/run-7y4j7voo.wandb differ diff --git a/wandb/run-20250914_045350-z1nqkm39/files/config.yaml b/wandb/run-20250914_045350-z1nqkm39/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7359091c32bf3231392aeb51388e1cd76df0b8df --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/files/config.yaml @@ -0,0 +1,114 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 9qty7bfmocoqc639x7jr9kqgva60fcyy: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.val_check_interval=0 + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533431480320" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T20:53:50.980969Z" + writerId: 9qty7bfmocoqc639x7jr9kqgva60fcyy + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_045350-z1nqkm39/files/output.log b/wandb/run-20250914_045350-z1nqkm39/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..2885e93a038c55739f708f539b2a629202c4a82a --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/files/output.log @@ -0,0 +1,106 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +Epoch 0: 50%|████████████████████▌ | 1/2 [00:22<00:22, 0.05it/s, v_num=km39] +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main + cli = LightningCLI(save_config_kwargs={"overwrite": True}) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 153, in run + self.on_advance_end(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 382, in on_advance_end + should_check_val = self._should_check_val_fx(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 545, in _should_check_val_fx + is_val_check_batch = (current_iteration + 1) % self.trainer.val_check_batch == 0 +ZeroDivisionError: integer division or modulo by zero +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 10, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 7, in cli_main +[rank0]: cli = LightningCLI(save_config_kwargs={"overwrite": True}) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 153, in run +[rank0]: self.on_advance_end(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 382, in on_advance_end +[rank0]: should_check_val = self._should_check_val_fx(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 545, in _should_check_val_fx +[rank0]: is_val_check_batch = (current_iteration + 1) % self.trainer.val_check_batch == 0 +[rank0]: ZeroDivisionError: integer division or modulo by zero diff --git a/wandb/run-20250914_045350-z1nqkm39/files/requirements.txt b/wandb/run-20250914_045350-z1nqkm39/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_045350-z1nqkm39/files/wandb-metadata.json b/wandb/run-20250914_045350-z1nqkm39/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a92dbc9d71646da3d89acba143c2492e3029ba9b --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/files/wandb-metadata.json @@ -0,0 +1,57 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T20:53:50.980969Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=0", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533431480320" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "9qty7bfmocoqc639x7jr9kqgva60fcyy" +} \ No newline at end of file diff --git a/wandb/run-20250914_045350-z1nqkm39/files/wandb-summary.json b/wandb/run-20250914_045350-z1nqkm39/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..76a77de4a5682ee7772a6a2c27390461e7c7fac9 --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":63,"_wandb":{"runtime":63}} \ No newline at end of file diff --git a/wandb/run-20250914_045350-z1nqkm39/logs/debug-core.log b/wandb/run-20250914_045350-z1nqkm39/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..ea91cf9347e0ca5b6ef89c2b834c4fcd0ecffb42 --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T04:53:51.048898571+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpj8uo8vg2/port-3710912.txt","pid":3710912,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T04:53:51.049160743+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T04:53:51.050600798+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3710912} +{"time":"2025-09-14T04:53:51.050562783+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3710912-3743527-2225806104/socket","Net":"unix"}} +{"time":"2025-09-14T04:53:51.239099748+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T04:53:51.255232392+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"z1nqkm39","id":"1(@)"} +{"time":"2025-09-14T04:53:51.719633015+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"z1nqkm39","id":"1(@)"} +{"time":"2025-09-14T04:54:55.540289058+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T04:54:55.540506374+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T04:54:55.540669656+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T04:54:55.540806927+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T04:54:55.541329859+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3710912-3743527-2225806104/socket","Net":"unix"}} +{"time":"2025-09-14T04:54:57.10225534+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T04:54:57.102331554+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T04:54:57.102353505+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_045350-z1nqkm39/logs/debug-internal.log b/wandb/run-20250914_045350-z1nqkm39/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..af7866380691da37066a37de825d24dbab0c9605 --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T04:53:51.255382613+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T04:53:51.274356476+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T04:53:51.719390639+08:00","level":"INFO","msg":"stream: created new stream","id":"z1nqkm39"} +{"time":"2025-09-14T04:53:51.719605864+08:00","level":"INFO","msg":"stream: started","id":"z1nqkm39"} +{"time":"2025-09-14T04:53:51.719653231+08:00","level":"INFO","msg":"writer: started","stream_id":"z1nqkm39"} +{"time":"2025-09-14T04:53:51.719800817+08:00","level":"INFO","msg":"sender: started","stream_id":"z1nqkm39"} +{"time":"2025-09-14T04:53:51.71990134+08:00","level":"INFO","msg":"handler: started","stream_id":"z1nqkm39"} +{"time":"2025-09-14T04:54:55.54049844+08:00","level":"INFO","msg":"stream: closing","id":"z1nqkm39"} +{"time":"2025-09-14T04:54:56.834345514+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T04:54:57.100177248+08:00","level":"INFO","msg":"handler: closed","stream_id":"z1nqkm39"} +{"time":"2025-09-14T04:54:57.101047274+08:00","level":"INFO","msg":"sender: closed","stream_id":"z1nqkm39"} +{"time":"2025-09-14T04:54:57.101138619+08:00","level":"INFO","msg":"stream: closed","id":"z1nqkm39"} diff --git a/wandb/run-20250914_045350-z1nqkm39/logs/debug.log b/wandb/run-20250914_045350-z1nqkm39/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..5f1a9472ca4ee51720862f7271aa7ff59f7ce1ff --- /dev/null +++ b/wandb/run-20250914_045350-z1nqkm39/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 04:53:50,988 INFO MainThread:3710912 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 04:53:50,988 INFO MainThread:3710912 [wandb_setup.py:_flush():81] Configure stats pid to 3710912 +2025-09-14 04:53:50,988 INFO MainThread:3710912 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 04:53:50,989 INFO MainThread:3710912 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 04:53:50,989 INFO MainThread:3710912 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 04:53:50,989 INFO MainThread:3710912 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_045350-z1nqkm39/logs/debug.log +2025-09-14 04:53:50,989 INFO MainThread:3710912 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_045350-z1nqkm39/logs/debug-internal.log +2025-09-14 04:53:50,990 INFO MainThread:3710912 [wandb_init.py:init():813] calling init triggers +2025-09-14 04:53:50,990 INFO MainThread:3710912 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 04:53:50,990 INFO MainThread:3710912 [wandb_init.py:init():854] starting backend +2025-09-14 04:53:51,240 INFO MainThread:3710912 [wandb_init.py:init():857] sending inform_init request +2025-09-14 04:53:51,252 INFO MainThread:3710912 [wandb_init.py:init():865] backend started and connected +2025-09-14 04:53:51,265 INFO MainThread:3710912 [wandb_init.py:init():936] updated telemetry +2025-09-14 04:53:51,289 INFO MainThread:3710912 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 04:53:52,034 INFO MainThread:3710912 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 04:53:52,495 INFO MainThread:3710912 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 04:53:52,496 INFO MainThread:3710912 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 04:53:52,497 INFO MainThread:3710912 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 04:53:52,497 INFO MainThread:3710912 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 04:53:52,509 INFO MainThread:3710912 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 04:54:25,120 INFO MainThread:3710912 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 04:54:55,540 INFO wandb-AsyncioManager-main:3710912 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 04:54:55,541 INFO wandb-AsyncioManager-main:3710912 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_045350-z1nqkm39/run-z1nqkm39.wandb b/wandb/run-20250914_045350-z1nqkm39/run-z1nqkm39.wandb new file mode 100644 index 0000000000000000000000000000000000000000..8878b2b6d0666901294c1c2da5639469393d6243 Binary files /dev/null and b/wandb/run-20250914_045350-z1nqkm39/run-z1nqkm39.wandb differ diff --git a/wandb/run-20250914_050205-gbu0tx1h/files/config.yaml b/wandb/run-20250914_050205-gbu0tx1h/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..87dafb0bd3dc5628c2e16f128d7ec42f5f57ecc2 --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/files/config.yaml @@ -0,0 +1,114 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + rvq0xpl3rw33j451810qgjt3pmuqxnxe: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.val_check_interval=0.0 + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533432283136" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T21:02:05.889913Z" + writerId: rvq0xpl3rw33j451810qgjt3pmuqxnxe + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_050205-gbu0tx1h/files/output.log b/wandb/run-20250914_050205-gbu0tx1h/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..08b6df08eb0abf12a769a6539707b7e7c820d5ad --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/files/output.log @@ -0,0 +1,12 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +Epoch 0: 100%|█████████████████████████████████████████████████████████| 2/2 [00:42<00:00, 0.05it/s, v_num=tx1h] +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. + +`Trainer.fit` stopped: `max_epochs=1` reached. diff --git a/wandb/run-20250914_050205-gbu0tx1h/files/requirements.txt b/wandb/run-20250914_050205-gbu0tx1h/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_050205-gbu0tx1h/files/wandb-metadata.json b/wandb/run-20250914_050205-gbu0tx1h/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c45c0e5435ba0393e2302865b3c78ae184f741fa --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/files/wandb-metadata.json @@ -0,0 +1,57 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T21:02:05.889913Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=0.0", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533432283136" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "rvq0xpl3rw33j451810qgjt3pmuqxnxe" +} \ No newline at end of file diff --git a/wandb/run-20250914_050205-gbu0tx1h/files/wandb-summary.json b/wandb/run-20250914_050205-gbu0tx1h/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..48f7d9a062ccb1f6a06780718d54c131168a079a --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":77},"_runtime":77} \ No newline at end of file diff --git a/wandb/run-20250914_050205-gbu0tx1h/logs/debug-core.log b/wandb/run-20250914_050205-gbu0tx1h/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..b7b8deaa28acc891ce03c864a058a9abc7fe8f88 --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T05:02:05.959684738+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp5s9_77d0/port-3758524.txt","pid":3758524,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T05:02:05.95986854+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T05:02:05.960551467+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3758524} +{"time":"2025-09-14T05:02:05.960545822+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3758524-3770088-2835303097/socket","Net":"unix"}} +{"time":"2025-09-14T05:02:06.151348602+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T05:02:06.169326276+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"gbu0tx1h","id":"1(@)"} +{"time":"2025-09-14T05:02:06.668529101+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"gbu0tx1h","id":"1(@)"} +{"time":"2025-09-14T05:03:24.094725697+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T05:03:24.094929478+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T05:03:24.094923519+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T05:03:24.0950962+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T05:03:24.095153138+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3758524-3770088-2835303097/socket","Net":"unix"}} +{"time":"2025-09-14T05:03:25.715959526+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T05:03:25.716025982+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T05:03:25.716048189+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_050205-gbu0tx1h/logs/debug-internal.log b/wandb/run-20250914_050205-gbu0tx1h/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..c072440b060f64df6c4089cdce725a30ccc31b25 --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T05:02:06.16951218+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T05:02:06.214583841+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T05:02:06.668392358+08:00","level":"INFO","msg":"stream: created new stream","id":"gbu0tx1h"} +{"time":"2025-09-14T05:02:06.668512429+08:00","level":"INFO","msg":"stream: started","id":"gbu0tx1h"} +{"time":"2025-09-14T05:02:06.668574907+08:00","level":"INFO","msg":"writer: started","stream_id":"gbu0tx1h"} +{"time":"2025-09-14T05:02:06.668601307+08:00","level":"INFO","msg":"sender: started","stream_id":"gbu0tx1h"} +{"time":"2025-09-14T05:02:06.668716538+08:00","level":"INFO","msg":"handler: started","stream_id":"gbu0tx1h"} +{"time":"2025-09-14T05:03:24.094882546+08:00","level":"INFO","msg":"stream: closing","id":"gbu0tx1h"} +{"time":"2025-09-14T05:03:25.342103303+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T05:03:25.714448533+08:00","level":"INFO","msg":"handler: closed","stream_id":"gbu0tx1h"} +{"time":"2025-09-14T05:03:25.714994341+08:00","level":"INFO","msg":"sender: closed","stream_id":"gbu0tx1h"} +{"time":"2025-09-14T05:03:25.715036148+08:00","level":"INFO","msg":"stream: closed","id":"gbu0tx1h"} diff --git a/wandb/run-20250914_050205-gbu0tx1h/logs/debug.log b/wandb/run-20250914_050205-gbu0tx1h/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..aba9fcada1d37133d28a52215545d88bf4974e29 --- /dev/null +++ b/wandb/run-20250914_050205-gbu0tx1h/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 05:02:05,897 INFO MainThread:3758524 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 05:02:05,897 INFO MainThread:3758524 [wandb_setup.py:_flush():81] Configure stats pid to 3758524 +2025-09-14 05:02:05,897 INFO MainThread:3758524 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 05:02:05,898 INFO MainThread:3758524 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 05:02:05,898 INFO MainThread:3758524 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 05:02:05,898 INFO MainThread:3758524 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_050205-gbu0tx1h/logs/debug.log +2025-09-14 05:02:05,899 INFO MainThread:3758524 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_050205-gbu0tx1h/logs/debug-internal.log +2025-09-14 05:02:05,899 INFO MainThread:3758524 [wandb_init.py:init():813] calling init triggers +2025-09-14 05:02:05,899 INFO MainThread:3758524 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 05:02:05,900 INFO MainThread:3758524 [wandb_init.py:init():854] starting backend +2025-09-14 05:02:06,153 INFO MainThread:3758524 [wandb_init.py:init():857] sending inform_init request +2025-09-14 05:02:06,165 INFO MainThread:3758524 [wandb_init.py:init():865] backend started and connected +2025-09-14 05:02:06,174 INFO MainThread:3758524 [wandb_init.py:init():936] updated telemetry +2025-09-14 05:02:06,200 INFO MainThread:3758524 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 05:02:06,944 INFO MainThread:3758524 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 05:02:07,362 INFO MainThread:3758524 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 05:02:07,363 INFO MainThread:3758524 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 05:02:07,364 INFO MainThread:3758524 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 05:02:07,364 INFO MainThread:3758524 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 05:02:07,371 INFO MainThread:3758524 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 05:02:38,582 INFO MainThread:3758524 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 05:03:24,095 INFO wandb-AsyncioManager-main:3758524 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 05:03:24,095 INFO wandb-AsyncioManager-main:3758524 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_050205-gbu0tx1h/run-gbu0tx1h.wandb b/wandb/run-20250914_050205-gbu0tx1h/run-gbu0tx1h.wandb new file mode 100644 index 0000000000000000000000000000000000000000..faf5327b63138842c99aca24dd98c1554d68d315 Binary files /dev/null and b/wandb/run-20250914_050205-gbu0tx1h/run-gbu0tx1h.wandb differ diff --git a/wandb/run-20250914_051233-guunipg5/files/config.yaml b/wandb/run-20250914_051233-guunipg5/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..39c1dcfb437120755b8238ae039e5f01c1d8620b --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/files/config.yaml @@ -0,0 +1,114 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + kvj1aqfhsn7sd24kig4si95nfpa81vaj: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.val_check_interval=0.0 + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533433274368" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T21:12:33.989020Z" + writerId: kvj1aqfhsn7sd24kig4si95nfpa81vaj + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_051233-guunipg5/files/output.log b/wandb/run-20250914_051233-guunipg5/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..238ff358bef21a3915c3667d1bee5c65eed87e84 --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/files/output.log @@ -0,0 +1,12 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +Epoch 0: 100%|████████████████████████████████| 2/2 [00:41<00:00, 0.05it/s, v_num=ipg5] +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Namespace(config=None, subcommand='fit', fit=Namespace(config=None, seed_everything=0, trainer=Namespace(accelerator='gpu', strategy='deepspeed_stage_3', devices='8', num_nodes=1, precision='bf16', logger=Namespace(class_path='lightning.pytorch.loggers.WandbLogger', init_args=Namespace(name='Test-AC-critic_expectile_0.9-inv_temp_1.0', save_dir='.', version=None, offline=False, dir=None, id=None, anonymous=None, project='RSAGame-Official', log_model=False, experiment=None, prefix='', checkpoint_name=None, entity=None, notes=None, tags=None, config=None, config_exclude_keys=None, config_include_keys=None, allow_val_change=None, group=None, job_type=None, force=None, reinit=None, resume=None, resume_from=None, fork_from=None, save_code=None, tensorboard=None, sync_tensorboard=None, monitor_gym=None, settings=None)), callbacks=None, fast_dev_run=False, max_epochs=1, min_epochs=None, max_steps=-1, min_steps=None, max_time=None, limit_train_batches=None, limit_val_batches=None, limit_test_batches=None, limit_predict_batches=None, overfit_batches=0.0, val_check_interval=0.0, check_val_every_n_epoch=1, num_sanity_val_steps=None, log_every_n_steps=None, enable_checkpointing=None, enable_progress_bar=None, enable_model_summary=False, accumulate_grad_batches=1, gradient_clip_val=None, gradient_clip_algorithm=None, deterministic=None, benchmark=None, inference_mode=True, use_distributed_sampler=True, profiler=None, detect_anomaly=False, barebones=False, plugins=None, sync_batchnorm=False, reload_dataloaders_every_n_epochs=0, default_root_dir='checkpoints/archer_Qwen3-14B_rsa', model_registry=None), model=Namespace(class_path='Algorithms.OfflineArcher', init_args=Namespace(model_name_or_path='/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', inv_temp=1.0, actor_lr=1e-05, critic_lr=1e-05, tau=0.05, accumulate_grad_batches=16, discount_factor=0.99, critic_expectile=0.9, optimize_critic=True, actor_checkpoint=None, critic_checkpoint=None)), data=Namespace(class_path='Tasks.RSAGame', init_args=Namespace(base_model='Qwen3-14B', batch_size=2, n_traj_eval=4, word_list=None)), optimizer=None, lr_scheduler=None, ckpt_path=None)) +`Trainer.fit` stopped: `max_epochs=1` reached. diff --git a/wandb/run-20250914_051233-guunipg5/files/requirements.txt b/wandb/run-20250914_051233-guunipg5/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_051233-guunipg5/files/wandb-metadata.json b/wandb/run-20250914_051233-guunipg5/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e15f97f519e4371a7ac0eff7bfea44cefc673f53 --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/files/wandb-metadata.json @@ -0,0 +1,57 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T21:12:33.989020Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=0.0", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533433274368" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "kvj1aqfhsn7sd24kig4si95nfpa81vaj" +} \ No newline at end of file diff --git a/wandb/run-20250914_051233-guunipg5/files/wandb-summary.json b/wandb/run-20250914_051233-guunipg5/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..022646b35b7c26b16cc2e065b40d7c01e5cfafde --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":76},"_runtime":76} \ No newline at end of file diff --git a/wandb/run-20250914_051233-guunipg5/logs/debug-core.log b/wandb/run-20250914_051233-guunipg5/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..e6bd9f371c3c6f44af02020f35e60493d309838a --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T05:12:34.055905556+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp0gt5b4rw/port-3791843.txt","pid":3791843,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T05:12:34.056105719+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T05:12:34.058174618+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3791843} +{"time":"2025-09-14T05:12:34.058146011+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3791843-3801899-1603454528/socket","Net":"unix"}} +{"time":"2025-09-14T05:12:34.244376318+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T05:12:34.258043783+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"guunipg5","id":"1(@)"} +{"time":"2025-09-14T05:12:34.726727836+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"guunipg5","id":"1(@)"} +{"time":"2025-09-14T05:13:51.170695419+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T05:13:51.170808512+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T05:13:51.17078854+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T05:13:51.171059537+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T05:13:51.171172933+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3791843-3801899-1603454528/socket","Net":"unix"}} +{"time":"2025-09-14T05:13:52.732204311+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T05:13:52.732249689+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T05:13:52.732268236+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_051233-guunipg5/logs/debug-internal.log b/wandb/run-20250914_051233-guunipg5/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..00c907e6902585306cf75bf0a93b679fd061f184 --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T05:12:34.258248259+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T05:12:34.287159572+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T05:12:34.726626235+08:00","level":"INFO","msg":"stream: created new stream","id":"guunipg5"} +{"time":"2025-09-14T05:12:34.726712029+08:00","level":"INFO","msg":"stream: started","id":"guunipg5"} +{"time":"2025-09-14T05:12:34.726742095+08:00","level":"INFO","msg":"writer: started","stream_id":"guunipg5"} +{"time":"2025-09-14T05:12:34.726762146+08:00","level":"INFO","msg":"handler: started","stream_id":"guunipg5"} +{"time":"2025-09-14T05:12:34.726885797+08:00","level":"INFO","msg":"sender: started","stream_id":"guunipg5"} +{"time":"2025-09-14T05:13:51.170790359+08:00","level":"INFO","msg":"stream: closing","id":"guunipg5"} +{"time":"2025-09-14T05:13:52.465094671+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T05:13:52.730433221+08:00","level":"INFO","msg":"handler: closed","stream_id":"guunipg5"} +{"time":"2025-09-14T05:13:52.731464685+08:00","level":"INFO","msg":"sender: closed","stream_id":"guunipg5"} +{"time":"2025-09-14T05:13:52.73152327+08:00","level":"INFO","msg":"stream: closed","id":"guunipg5"} diff --git a/wandb/run-20250914_051233-guunipg5/logs/debug.log b/wandb/run-20250914_051233-guunipg5/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..132322b61eb2bb424c078e645e3622d430d29059 --- /dev/null +++ b/wandb/run-20250914_051233-guunipg5/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 05:12:33,999 INFO MainThread:3791843 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 05:12:33,999 INFO MainThread:3791843 [wandb_setup.py:_flush():81] Configure stats pid to 3791843 +2025-09-14 05:12:34,000 INFO MainThread:3791843 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 05:12:34,000 INFO MainThread:3791843 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 05:12:34,000 INFO MainThread:3791843 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 05:12:34,001 INFO MainThread:3791843 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_051233-guunipg5/logs/debug.log +2025-09-14 05:12:34,001 INFO MainThread:3791843 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_051233-guunipg5/logs/debug-internal.log +2025-09-14 05:12:34,001 INFO MainThread:3791843 [wandb_init.py:init():813] calling init triggers +2025-09-14 05:12:34,002 INFO MainThread:3791843 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 05:12:34,002 INFO MainThread:3791843 [wandb_init.py:init():854] starting backend +2025-09-14 05:12:34,245 INFO MainThread:3791843 [wandb_init.py:init():857] sending inform_init request +2025-09-14 05:12:34,253 INFO MainThread:3791843 [wandb_init.py:init():865] backend started and connected +2025-09-14 05:12:34,262 INFO MainThread:3791843 [wandb_init.py:init():936] updated telemetry +2025-09-14 05:12:34,286 INFO MainThread:3791843 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 05:12:35,067 INFO MainThread:3791843 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 05:12:35,501 INFO MainThread:3791843 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 05:12:35,502 INFO MainThread:3791843 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 05:12:35,503 INFO MainThread:3791843 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 05:12:35,503 INFO MainThread:3791843 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 05:12:35,512 INFO MainThread:3791843 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 05:13:06,887 INFO MainThread:3791843 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 05:13:51,170 INFO wandb-AsyncioManager-main:3791843 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 05:13:51,171 INFO wandb-AsyncioManager-main:3791843 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_051233-guunipg5/run-guunipg5.wandb b/wandb/run-20250914_051233-guunipg5/run-guunipg5.wandb new file mode 100644 index 0000000000000000000000000000000000000000..b8ff2b0152f028a0870c8f1533e69d4a9cef15d6 Binary files /dev/null and b/wandb/run-20250914_051233-guunipg5/run-guunipg5.wandb differ diff --git a/wandb/run-20250914_052413-mpyyfd9m/files/config.yaml b/wandb/run-20250914_052413-mpyyfd9m/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..104adfa4e699af4c4b6cca8592fb560672876303 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/files/config.yaml @@ -0,0 +1,114 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + oiinsnzwilppjczo1trk48vvxxpr7v76: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.val_check_interval=0.0 + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533434114048" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T21:24:13.102165Z" + writerId: oiinsnzwilppjczo1trk48vvxxpr7v76 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_052413-mpyyfd9m/files/output.log b/wandb/run-20250914_052413-mpyyfd9m/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b5a74ad6ad695f6d101b9369bffeb5c7cc3eda18 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/files/output.log @@ -0,0 +1,128 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +Epoch 0: 100%|████████████████████████████████| 2/2 [00:41<00:00, 0.05it/s, v_num=fd9m] +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. + +`Trainer.fit` stopped: `max_epochs=1` reached. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 21, in + if __name__ == "__main__": + File "/home/jiashuo/codes/OfflineArcher/main.py", line 12, in cli_main + ) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 223, in run + self.on_run_end() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 499, in on_run_end + call._call_callback_hooks(trainer, "on_train_end") + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks + fn(trainer, trainer.lightning_module, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 418, in on_train_end + self._save_last_checkpoint(trainer, monitor_candidates) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 793, in _save_last_checkpoint + self._save_checkpoint(trainer, filepath) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 473, in _save_checkpoint + trainer.save_checkpoint(filepath, self.save_weights_only) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1395, in save_checkpoint + checkpoint = self._checkpoint_connector.dump_checkpoint(weights_only) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/checkpoint_connector.py", line 502, in dump_checkpoint + call._call_lightning_module_hook(trainer, "on_save_checkpoint", checkpoint) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 177, in _call_lightning_module_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 609, in on_save_checkpoint + save_dir = self.trainer.checkpoint_callback.dirpath +NameError: name 'os' is not defined +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 21, in +[rank0]: if __name__ == "__main__": +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 12, in cli_main +[rank0]: ) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 223, in run +[rank0]: self.on_run_end() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 499, in on_run_end +[rank0]: call._call_callback_hooks(trainer, "on_train_end") +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks +[rank0]: fn(trainer, trainer.lightning_module, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 418, in on_train_end +[rank0]: self._save_last_checkpoint(trainer, monitor_candidates) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 793, in _save_last_checkpoint +[rank0]: self._save_checkpoint(trainer, filepath) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 473, in _save_checkpoint +[rank0]: trainer.save_checkpoint(filepath, self.save_weights_only) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1395, in save_checkpoint +[rank0]: checkpoint = self._checkpoint_connector.dump_checkpoint(weights_only) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/checkpoint_connector.py", line 502, in dump_checkpoint +[rank0]: call._call_lightning_module_hook(trainer, "on_save_checkpoint", checkpoint) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 177, in _call_lightning_module_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 609, in on_save_checkpoint +[rank0]: save_dir = self.trainer.checkpoint_callback.dirpath +[rank0]: NameError: name 'os' is not defined diff --git a/wandb/run-20250914_052413-mpyyfd9m/files/requirements.txt b/wandb/run-20250914_052413-mpyyfd9m/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_052413-mpyyfd9m/files/wandb-metadata.json b/wandb/run-20250914_052413-mpyyfd9m/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..18d69972b8cf2376d6917b0ee7a56b68e7f7bdd9 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/files/wandb-metadata.json @@ -0,0 +1,57 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T21:24:13.102165Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=0.0", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533434114048" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "oiinsnzwilppjczo1trk48vvxxpr7v76" +} \ No newline at end of file diff --git a/wandb/run-20250914_052413-mpyyfd9m/files/wandb-summary.json b/wandb/run-20250914_052413-mpyyfd9m/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..f91d215151de9697bff56965060551af6b6de802 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":108},"_runtime":108} \ No newline at end of file diff --git a/wandb/run-20250914_052413-mpyyfd9m/logs/debug-core.log b/wandb/run-20250914_052413-mpyyfd9m/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..564853eb0f269e05c4b78f73bf6f331fc44f12b7 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T05:24:13.175979343+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp0rwk0061/port-3821868.txt","pid":3821868,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T05:24:13.176224908+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T05:24:13.177037328+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3821868} +{"time":"2025-09-14T05:24:13.177043711+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3821868-3833135-3062534520/socket","Net":"unix"}} +{"time":"2025-09-14T05:24:13.367141046+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T05:24:13.383325363+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"mpyyfd9m","id":"1(@)"} +{"time":"2025-09-14T05:24:13.836764018+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"mpyyfd9m","id":"1(@)"} +{"time":"2025-09-14T05:26:02.792356498+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T05:26:02.792697266+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T05:26:02.792772569+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T05:26:02.792921871+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T05:26:02.79357102+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3821868-3833135-3062534520/socket","Net":"unix"}} +{"time":"2025-09-14T05:26:04.424707193+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T05:26:04.424762394+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T05:26:04.424788224+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_052413-mpyyfd9m/logs/debug-internal.log b/wandb/run-20250914_052413-mpyyfd9m/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d0eedd3df9b65b584c9d384c0738e081ea3db9e4 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T05:24:13.383452363+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T05:24:13.402561117+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T05:24:13.836626754+08:00","level":"INFO","msg":"stream: created new stream","id":"mpyyfd9m"} +{"time":"2025-09-14T05:24:13.836749185+08:00","level":"INFO","msg":"stream: started","id":"mpyyfd9m"} +{"time":"2025-09-14T05:24:13.836782819+08:00","level":"INFO","msg":"writer: started","stream_id":"mpyyfd9m"} +{"time":"2025-09-14T05:24:13.836797249+08:00","level":"INFO","msg":"handler: started","stream_id":"mpyyfd9m"} +{"time":"2025-09-14T05:24:13.836857845+08:00","level":"INFO","msg":"sender: started","stream_id":"mpyyfd9m"} +{"time":"2025-09-14T05:26:02.792558688+08:00","level":"INFO","msg":"stream: closing","id":"mpyyfd9m"} +{"time":"2025-09-14T05:26:04.133800831+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T05:26:04.422978002+08:00","level":"INFO","msg":"handler: closed","stream_id":"mpyyfd9m"} +{"time":"2025-09-14T05:26:04.423820341+08:00","level":"INFO","msg":"sender: closed","stream_id":"mpyyfd9m"} +{"time":"2025-09-14T05:26:04.423871238+08:00","level":"INFO","msg":"stream: closed","id":"mpyyfd9m"} diff --git a/wandb/run-20250914_052413-mpyyfd9m/logs/debug.log b/wandb/run-20250914_052413-mpyyfd9m/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..34dda90eeb5bb52f406d16a2290bc7717f786523 --- /dev/null +++ b/wandb/run-20250914_052413-mpyyfd9m/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 05:24:13,111 INFO MainThread:3821868 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 05:24:13,112 INFO MainThread:3821868 [wandb_setup.py:_flush():81] Configure stats pid to 3821868 +2025-09-14 05:24:13,112 INFO MainThread:3821868 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 05:24:13,112 INFO MainThread:3821868 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 05:24:13,112 INFO MainThread:3821868 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 05:24:13,113 INFO MainThread:3821868 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_052413-mpyyfd9m/logs/debug.log +2025-09-14 05:24:13,113 INFO MainThread:3821868 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_052413-mpyyfd9m/logs/debug-internal.log +2025-09-14 05:24:13,114 INFO MainThread:3821868 [wandb_init.py:init():813] calling init triggers +2025-09-14 05:24:13,114 INFO MainThread:3821868 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 05:24:13,114 INFO MainThread:3821868 [wandb_init.py:init():854] starting backend +2025-09-14 05:24:13,368 INFO MainThread:3821868 [wandb_init.py:init():857] sending inform_init request +2025-09-14 05:24:13,380 INFO MainThread:3821868 [wandb_init.py:init():865] backend started and connected +2025-09-14 05:24:13,394 INFO MainThread:3821868 [wandb_init.py:init():936] updated telemetry +2025-09-14 05:24:13,422 INFO MainThread:3821868 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 05:24:14,153 INFO MainThread:3821868 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 05:24:14,598 INFO MainThread:3821868 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 05:24:14,599 INFO MainThread:3821868 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 05:24:14,600 INFO MainThread:3821868 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 05:24:14,600 INFO MainThread:3821868 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 05:24:14,609 INFO MainThread:3821868 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 05:24:46,778 INFO MainThread:3821868 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 05:26:02,793 INFO wandb-AsyncioManager-main:3821868 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 05:26:02,793 INFO wandb-AsyncioManager-main:3821868 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_052413-mpyyfd9m/run-mpyyfd9m.wandb b/wandb/run-20250914_052413-mpyyfd9m/run-mpyyfd9m.wandb new file mode 100644 index 0000000000000000000000000000000000000000..18ae39ffc146f40f6697edd5ed24d6e48b9691cb Binary files /dev/null and b/wandb/run-20250914_052413-mpyyfd9m/run-mpyyfd9m.wandb differ diff --git a/wandb/run-20250914_053337-q8rjmhdj/files/config.yaml b/wandb/run-20250914_053337-q8rjmhdj/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ec332174f5e8702b1e8a305ed284f31e537a6e5c --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/files/config.yaml @@ -0,0 +1,114 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + pso9y0jtbxacg0s9ozoh0gcw3tj8x2fp: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.val_check_interval=0.0 + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533435056128" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T21:33:37.767286Z" + writerId: pso9y0jtbxacg0s9ozoh0gcw3tj8x2fp + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_053337-q8rjmhdj/files/output.log b/wandb/run-20250914_053337-q8rjmhdj/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ee5df3534e1d554a24faa0b5bb874f4907fb9f29 --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/files/output.log @@ -0,0 +1,146 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +Epoch 0: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [00:41<00:00, 0.05it/s, v_num=mhdj] +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'val_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +✅ LoRA adapter saved to: /home/jiashuo/codes/OfflineArcher/${trainer.default_root_dir} +`Trainer.fit` stopped: `max_epochs=1` reached. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 223, in run + self.on_run_end() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 499, in on_run_end + call._call_callback_hooks(trainer, "on_train_end") + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks + fn(trainer, trainer.lightning_module, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 418, in on_train_end + self._save_last_checkpoint(trainer, monitor_candidates) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 793, in _save_last_checkpoint + self._save_checkpoint(trainer, filepath) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 473, in _save_checkpoint + trainer.save_checkpoint(filepath, self.save_weights_only) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1395, in save_checkpoint + checkpoint = self._checkpoint_connector.dump_checkpoint(weights_only) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/checkpoint_connector.py", line 502, in dump_checkpoint + call._call_lightning_module_hook(trainer, "on_save_checkpoint", checkpoint) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 177, in _call_lightning_module_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 621, in on_save_checkpoint + self.merge_and_save_lora(os.path.join(save_dir, "merged_model")) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 629, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 678, in merge + delta_weight = self.get_delta_weight(active_adapter) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 733, in get_delta_weight + output_tensor = transpose(weight_B @ weight_A, self.fan_in_fan_out) * self.scaling[adapter] +RuntimeError: mat1 and mat2 shapes cannot be multiplied (1x0 and 8x5120) +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 223, in run +[rank0]: self.on_run_end() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 499, in on_run_end +[rank0]: call._call_callback_hooks(trainer, "on_train_end") +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks +[rank0]: fn(trainer, trainer.lightning_module, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 418, in on_train_end +[rank0]: self._save_last_checkpoint(trainer, monitor_candidates) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 793, in _save_last_checkpoint +[rank0]: self._save_checkpoint(trainer, filepath) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 473, in _save_checkpoint +[rank0]: trainer.save_checkpoint(filepath, self.save_weights_only) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1395, in save_checkpoint +[rank0]: checkpoint = self._checkpoint_connector.dump_checkpoint(weights_only) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/checkpoint_connector.py", line 502, in dump_checkpoint +[rank0]: call._call_lightning_module_hook(trainer, "on_save_checkpoint", checkpoint) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 177, in _call_lightning_module_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 621, in on_save_checkpoint +[rank0]: self.merge_and_save_lora(os.path.join(save_dir, "merged_model")) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 629, in merge_and_save_lora +[rank0]: merged_model = self.actor.model.merge_and_unload() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload +[rank0]: return self._unload_and_optionally_merge( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge +[rank0]: target.merge(safe_merge=safe_merge, adapter_names=adapter_names) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 678, in merge +[rank0]: delta_weight = self.get_delta_weight(active_adapter) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 733, in get_delta_weight +[rank0]: output_tensor = transpose(weight_B @ weight_A, self.fan_in_fan_out) * self.scaling[adapter] +[rank0]: RuntimeError: mat1 and mat2 shapes cannot be multiplied (1x0 and 8x5120) diff --git a/wandb/run-20250914_053337-q8rjmhdj/files/requirements.txt b/wandb/run-20250914_053337-q8rjmhdj/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_053337-q8rjmhdj/files/wandb-metadata.json b/wandb/run-20250914_053337-q8rjmhdj/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..d4d87126dd79f2e53d86a40911054763d961074b --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/files/wandb-metadata.json @@ -0,0 +1,57 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T21:33:37.767286Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.val_check_interval=0.0", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533435056128" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "pso9y0jtbxacg0s9ozoh0gcw3tj8x2fp" +} \ No newline at end of file diff --git a/wandb/run-20250914_053337-q8rjmhdj/files/wandb-summary.json b/wandb/run-20250914_053337-q8rjmhdj/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..2117feacff368688889a165c0ee250b1b0ca2970 --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":187},"_runtime":187} \ No newline at end of file diff --git a/wandb/run-20250914_053337-q8rjmhdj/logs/debug-core.log b/wandb/run-20250914_053337-q8rjmhdj/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..4df87e8e76b0b5573dd6265e08a0eb0574c67cfa --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T05:33:37.834376785+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpwey9ydcu/port-3854863.txt","pid":3854863,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T05:33:37.834682832+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T05:33:37.835630526+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3854863} +{"time":"2025-09-14T05:33:37.835603115+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3854863-3863792-2992393324/socket","Net":"unix"}} +{"time":"2025-09-14T05:33:38.025286179+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T05:33:38.051142395+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"q8rjmhdj","id":"1(@)"} +{"time":"2025-09-14T05:33:38.507639304+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"q8rjmhdj","id":"1(@)"} +{"time":"2025-09-14T05:36:45.932463456+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T05:36:45.93274383+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T05:36:45.932894131+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T05:36:45.932966142+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T05:36:45.933393724+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3854863-3863792-2992393324/socket","Net":"unix"}} +{"time":"2025-09-14T05:36:47.851325512+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T05:36:47.85135919+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T05:36:47.851369067+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_053337-q8rjmhdj/logs/debug-internal.log b/wandb/run-20250914_053337-q8rjmhdj/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a83211d8749493cfa40e9d662ab908b7c229ab31 --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T05:33:38.051315202+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T05:33:38.070000848+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T05:33:38.507523318+08:00","level":"INFO","msg":"stream: created new stream","id":"q8rjmhdj"} +{"time":"2025-09-14T05:33:38.507624486+08:00","level":"INFO","msg":"stream: started","id":"q8rjmhdj"} +{"time":"2025-09-14T05:33:38.507656011+08:00","level":"INFO","msg":"writer: started","stream_id":"q8rjmhdj"} +{"time":"2025-09-14T05:33:38.507708681+08:00","level":"INFO","msg":"sender: started","stream_id":"q8rjmhdj"} +{"time":"2025-09-14T05:33:38.507698311+08:00","level":"INFO","msg":"handler: started","stream_id":"q8rjmhdj"} +{"time":"2025-09-14T05:36:45.932651819+08:00","level":"INFO","msg":"stream: closing","id":"q8rjmhdj"} +{"time":"2025-09-14T05:36:47.227805569+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T05:36:47.84961516+08:00","level":"INFO","msg":"handler: closed","stream_id":"q8rjmhdj"} +{"time":"2025-09-14T05:36:47.850273923+08:00","level":"INFO","msg":"sender: closed","stream_id":"q8rjmhdj"} +{"time":"2025-09-14T05:36:47.85031413+08:00","level":"INFO","msg":"stream: closed","id":"q8rjmhdj"} diff --git a/wandb/run-20250914_053337-q8rjmhdj/logs/debug.log b/wandb/run-20250914_053337-q8rjmhdj/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..8b8e52e4f26cf680048bae5234ffafc9e4502d55 --- /dev/null +++ b/wandb/run-20250914_053337-q8rjmhdj/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 05:33:37,773 INFO MainThread:3854863 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 05:33:37,774 INFO MainThread:3854863 [wandb_setup.py:_flush():81] Configure stats pid to 3854863 +2025-09-14 05:33:37,774 INFO MainThread:3854863 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 05:33:37,775 INFO MainThread:3854863 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 05:33:37,775 INFO MainThread:3854863 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 05:33:37,775 INFO MainThread:3854863 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_053337-q8rjmhdj/logs/debug.log +2025-09-14 05:33:37,776 INFO MainThread:3854863 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_053337-q8rjmhdj/logs/debug-internal.log +2025-09-14 05:33:37,776 INFO MainThread:3854863 [wandb_init.py:init():813] calling init triggers +2025-09-14 05:33:37,776 INFO MainThread:3854863 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 05:33:37,777 INFO MainThread:3854863 [wandb_init.py:init():854] starting backend +2025-09-14 05:33:38,026 INFO MainThread:3854863 [wandb_init.py:init():857] sending inform_init request +2025-09-14 05:33:38,036 INFO MainThread:3854863 [wandb_init.py:init():865] backend started and connected +2025-09-14 05:33:38,045 INFO MainThread:3854863 [wandb_init.py:init():936] updated telemetry +2025-09-14 05:33:38,070 INFO MainThread:3854863 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 05:33:38,880 INFO MainThread:3854863 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 05:33:39,311 INFO MainThread:3854863 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 05:33:39,312 INFO MainThread:3854863 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 05:33:39,313 INFO MainThread:3854863 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 05:33:39,313 INFO MainThread:3854863 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 05:33:39,319 INFO MainThread:3854863 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 05:34:09,466 INFO MainThread:3854863 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 05:36:45,933 INFO wandb-AsyncioManager-main:3854863 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 05:36:45,933 INFO wandb-AsyncioManager-main:3854863 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_053337-q8rjmhdj/run-q8rjmhdj.wandb b/wandb/run-20250914_053337-q8rjmhdj/run-q8rjmhdj.wandb new file mode 100644 index 0000000000000000000000000000000000000000..53f79fec9baa9ccfa5ccb9050addb0e2ff0c8054 Binary files /dev/null and b/wandb/run-20250914_053337-q8rjmhdj/run-q8rjmhdj.wandb differ diff --git a/wandb/run-20250914_054327-76ohsw1w/files/output.log b/wandb/run-20250914_054327-76ohsw1w/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..5b91fab0dc3fd5eecb1dbd5a387299d9c3e59893 --- /dev/null +++ b/wandb/run-20250914_054327-76ohsw1w/files/output.log @@ -0,0 +1,5 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +[rank: 1] Child process with PID 3884847 terminated with code 1. Forcefully terminating all other processes to avoid zombies 🧟 diff --git a/wandb/run-20250914_054327-76ohsw1w/files/requirements.txt b/wandb/run-20250914_054327-76ohsw1w/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_054327-76ohsw1w/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_054327-76ohsw1w/files/wandb-metadata.json b/wandb/run-20250914_054327-76ohsw1w/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c4525905573c1f32eddbdeb39847203889d760a8 --- /dev/null +++ b/wandb/run-20250914_054327-76ohsw1w/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T21:43:27.900599Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533410775040" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "w9g2j7ttu4eoqv0uzpox2usaevlda6w6" +} \ No newline at end of file diff --git a/wandb/run-20250914_054327-76ohsw1w/logs/debug-core.log b/wandb/run-20250914_054327-76ohsw1w/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..60ea90ffa1582e8f9a2e715a2ae1c45db3fa7779 --- /dev/null +++ b/wandb/run-20250914_054327-76ohsw1w/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T05:43:27.966939622+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmplp95_foq/port-3884551.txt","pid":3884551,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T05:43:27.967117225+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T05:43:27.967651918+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3884551} +{"time":"2025-09-14T05:43:27.967653749+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3884551-3890425-659378609/socket","Net":"unix"}} +{"time":"2025-09-14T05:43:28.157777338+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T05:43:28.173892365+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"76ohsw1w","id":"1(@)"} +{"time":"2025-09-14T05:43:28.673618895+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"76ohsw1w","id":"1(@)"} +{"time":"2025-09-14T05:47:19.806590658+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_054327-76ohsw1w/logs/debug-internal.log b/wandb/run-20250914_054327-76ohsw1w/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..6ca747373510a42311bbe3dd14294fb2147b6163 --- /dev/null +++ b/wandb/run-20250914_054327-76ohsw1w/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T05:43:28.174166222+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T05:43:28.21902985+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T05:43:28.673517569+08:00","level":"INFO","msg":"stream: created new stream","id":"76ohsw1w"} +{"time":"2025-09-14T05:43:28.673604287+08:00","level":"INFO","msg":"stream: started","id":"76ohsw1w"} +{"time":"2025-09-14T05:43:28.673632305+08:00","level":"INFO","msg":"writer: started","stream_id":"76ohsw1w"} +{"time":"2025-09-14T05:43:28.673711068+08:00","level":"INFO","msg":"sender: started","stream_id":"76ohsw1w"} +{"time":"2025-09-14T05:43:28.67365192+08:00","level":"INFO","msg":"handler: started","stream_id":"76ohsw1w"} diff --git a/wandb/run-20250914_054327-76ohsw1w/logs/debug.log b/wandb/run-20250914_054327-76ohsw1w/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..46a28e53ba820a7f138abec1aa20c9b68ededbcb --- /dev/null +++ b/wandb/run-20250914_054327-76ohsw1w/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 05:43:27,907 INFO MainThread:3884551 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 05:43:27,907 INFO MainThread:3884551 [wandb_setup.py:_flush():81] Configure stats pid to 3884551 +2025-09-14 05:43:27,908 INFO MainThread:3884551 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 05:43:27,908 INFO MainThread:3884551 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 05:43:27,908 INFO MainThread:3884551 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 05:43:27,909 INFO MainThread:3884551 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_054327-76ohsw1w/logs/debug.log +2025-09-14 05:43:27,909 INFO MainThread:3884551 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_054327-76ohsw1w/logs/debug-internal.log +2025-09-14 05:43:27,909 INFO MainThread:3884551 [wandb_init.py:init():813] calling init triggers +2025-09-14 05:43:27,910 INFO MainThread:3884551 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 05:43:27,910 INFO MainThread:3884551 [wandb_init.py:init():854] starting backend +2025-09-14 05:43:28,159 INFO MainThread:3884551 [wandb_init.py:init():857] sending inform_init request +2025-09-14 05:43:28,170 INFO MainThread:3884551 [wandb_init.py:init():865] backend started and connected +2025-09-14 05:43:28,177 INFO MainThread:3884551 [wandb_init.py:init():936] updated telemetry +2025-09-14 05:43:28,202 INFO MainThread:3884551 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 05:43:28,954 INFO MainThread:3884551 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 05:43:29,393 INFO MainThread:3884551 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 05:43:29,393 INFO MainThread:3884551 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 05:43:29,394 INFO MainThread:3884551 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 05:43:29,394 INFO MainThread:3884551 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 05:43:29,400 INFO MainThread:3884551 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_054327-76ohsw1w/run-76ohsw1w.wandb b/wandb/run-20250914_054327-76ohsw1w/run-76ohsw1w.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_055213-77omz43z/files/config.yaml b/wandb/run-20250914_055213-77omz43z/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e49b754cccf886ab313ef9305b048525d2423bed --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 0n59fqlem8hzydlkq8jsqfxwni2cyqxc: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.limit_val_batches=0 + - --trainer.val_check_interval=null + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3533411602432" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T21:52:13.130995Z" + writerId: 0n59fqlem8hzydlkq8jsqfxwni2cyqxc + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_055213-77omz43z/files/output.log b/wandb/run-20250914_055213-77omz43z/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..40cd07f71c1aec7fa982993feb35204cc5c88c92 --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/files/output.log @@ -0,0 +1,143 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 100%|█████████████████████████████████████████████████████████████████████████████████████████████| 2/2 [01:54<00:00, 0.02it/s, v_num=z43z] +`Trainer.fit` stopped: `max_epochs=1` reached. +✅ LoRA adapter saved to: checkpoints/archer_Qwen3-14B_rsa +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + if __name__ == "__main__": + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 223, in run + self.on_run_end() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 499, in on_run_end + call._call_callback_hooks(trainer, "on_train_end") + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks + fn(trainer, trainer.lightning_module, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 418, in on_train_end + self._save_last_checkpoint(trainer, monitor_candidates) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 793, in _save_last_checkpoint + self._save_checkpoint(trainer, filepath) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 473, in _save_checkpoint + trainer.save_checkpoint(filepath, self.save_weights_only) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1395, in save_checkpoint + checkpoint = self._checkpoint_connector.dump_checkpoint(weights_only) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/checkpoint_connector.py", line 502, in dump_checkpoint + call._call_lightning_module_hook(trainer, "on_save_checkpoint", checkpoint) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 177, in _call_lightning_module_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 621, in on_save_checkpoint + print(f"✅ LoRA adapter saved to: {save_dir}") + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 629, in merge_and_save_lora + if hasattr(self.actor.model, "merge_and_unload"): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (5120) at non-singleton dimension 1 +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main +[rank0]: return _run_code(code, main_globals, None, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in +[rank0]: cli.main() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main +[rank0]: run() +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file +[rank0]: runpy.run_path(target, run_name="__main__") +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path +[rank0]: return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code +[rank0]: _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) +[rank0]: File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code +[rank0]: exec(code, run_globals) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: if __name__ == "__main__": +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 223, in run +[rank0]: self.on_run_end() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 499, in on_run_end +[rank0]: call._call_callback_hooks(trainer, "on_train_end") +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 228, in _call_callback_hooks +[rank0]: fn(trainer, trainer.lightning_module, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 418, in on_train_end +[rank0]: self._save_last_checkpoint(trainer, monitor_candidates) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 793, in _save_last_checkpoint +[rank0]: self._save_checkpoint(trainer, filepath) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py", line 473, in _save_checkpoint +[rank0]: trainer.save_checkpoint(filepath, self.save_weights_only) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1395, in save_checkpoint +[rank0]: checkpoint = self._checkpoint_connector.dump_checkpoint(weights_only) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/checkpoint_connector.py", line 502, in dump_checkpoint +[rank0]: call._call_lightning_module_hook(trainer, "on_save_checkpoint", checkpoint) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 177, in _call_lightning_module_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 621, in on_save_checkpoint +[rank0]: print(f"✅ LoRA adapter saved to: {save_dir}") +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 629, in merge_and_save_lora +[rank0]: if hasattr(self.actor.model, "merge_and_unload"): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload +[rank0]: return self._unload_and_optionally_merge( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge +[rank0]: target.merge(safe_merge=safe_merge, adapter_names=adapter_names) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge +[rank0]: base_layer.weight.data += delta_weight +[rank0]: RuntimeError: The size of tensor a (0) must match the size of tensor b (5120) at non-singleton dimension 1 diff --git a/wandb/run-20250914_055213-77omz43z/files/requirements.txt b/wandb/run-20250914_055213-77omz43z/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_055213-77omz43z/files/wandb-metadata.json b/wandb/run-20250914_055213-77omz43z/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..08cb70161a04877e19c6da6dc40ff53748efbdd5 --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T21:52:13.130995Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533411602432" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "0n59fqlem8hzydlkq8jsqfxwni2cyqxc" +} \ No newline at end of file diff --git a/wandb/run-20250914_055213-77omz43z/files/wandb-summary.json b/wandb/run-20250914_055213-77omz43z/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..06c081eddf266e2eb5d941ac50f1117fa3d76e16 --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":268},"_runtime":268} \ No newline at end of file diff --git a/wandb/run-20250914_055213-77omz43z/logs/debug-core.log b/wandb/run-20250914_055213-77omz43z/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..e6e09afeeeb81801c9da191ace3501e7d062c810 --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T05:52:13.198649991+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp8zsp_d2k/port-3903645.txt","pid":3903645,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T05:52:13.198817853+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T05:52:13.199717643+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3903645} +{"time":"2025-09-14T05:52:13.199713236+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3903645-3913401-44537883/socket","Net":"unix"}} +{"time":"2025-09-14T05:52:13.390757903+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T05:52:13.407578538+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"77omz43z","id":"1(@)"} +{"time":"2025-09-14T05:52:13.900273741+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"77omz43z","id":"1(@)"} +{"time":"2025-09-14T05:56:43.090691719+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T05:56:43.091016647+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T05:56:43.091118083+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T05:56:43.09124+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T05:56:43.091777771+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3903645-3913401-44537883/socket","Net":"unix"}} +{"time":"2025-09-14T05:56:44.634105942+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T05:56:44.634171229+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T05:56:44.634214542+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_055213-77omz43z/logs/debug-internal.log b/wandb/run-20250914_055213-77omz43z/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..c3a40307edb4a520f54eabb6a97fd5bb9d25fb48 --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T05:52:13.407864567+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T05:52:13.453207428+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T05:52:13.900120429+08:00","level":"INFO","msg":"stream: created new stream","id":"77omz43z"} +{"time":"2025-09-14T05:52:13.900254786+08:00","level":"INFO","msg":"stream: started","id":"77omz43z"} +{"time":"2025-09-14T05:52:13.900343562+08:00","level":"INFO","msg":"writer: started","stream_id":"77omz43z"} +{"time":"2025-09-14T05:52:13.900373031+08:00","level":"INFO","msg":"sender: started","stream_id":"77omz43z"} +{"time":"2025-09-14T05:52:13.900478128+08:00","level":"INFO","msg":"handler: started","stream_id":"77omz43z"} +{"time":"2025-09-14T05:56:43.090889025+08:00","level":"INFO","msg":"stream: closing","id":"77omz43z"} +{"time":"2025-09-14T05:56:44.387727175+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T05:56:44.631898244+08:00","level":"INFO","msg":"handler: closed","stream_id":"77omz43z"} +{"time":"2025-09-14T05:56:44.632927829+08:00","level":"INFO","msg":"sender: closed","stream_id":"77omz43z"} +{"time":"2025-09-14T05:56:44.632986604+08:00","level":"INFO","msg":"stream: closed","id":"77omz43z"} diff --git a/wandb/run-20250914_055213-77omz43z/logs/debug.log b/wandb/run-20250914_055213-77omz43z/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..ccd55659c4560687281f776e37231e8b1f425af5 --- /dev/null +++ b/wandb/run-20250914_055213-77omz43z/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 05:52:13,136 INFO MainThread:3903645 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 05:52:13,137 INFO MainThread:3903645 [wandb_setup.py:_flush():81] Configure stats pid to 3903645 +2025-09-14 05:52:13,137 INFO MainThread:3903645 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 05:52:13,137 INFO MainThread:3903645 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 05:52:13,138 INFO MainThread:3903645 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 05:52:13,138 INFO MainThread:3903645 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_055213-77omz43z/logs/debug.log +2025-09-14 05:52:13,139 INFO MainThread:3903645 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_055213-77omz43z/logs/debug-internal.log +2025-09-14 05:52:13,139 INFO MainThread:3903645 [wandb_init.py:init():813] calling init triggers +2025-09-14 05:52:13,139 INFO MainThread:3903645 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 05:52:13,140 INFO MainThread:3903645 [wandb_init.py:init():854] starting backend +2025-09-14 05:52:13,392 INFO MainThread:3903645 [wandb_init.py:init():857] sending inform_init request +2025-09-14 05:52:13,402 INFO MainThread:3903645 [wandb_init.py:init():865] backend started and connected +2025-09-14 05:52:13,410 INFO MainThread:3903645 [wandb_init.py:init():936] updated telemetry +2025-09-14 05:52:13,436 INFO MainThread:3903645 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 05:52:14,261 INFO MainThread:3903645 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 05:52:14,697 INFO MainThread:3903645 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 05:52:14,697 INFO MainThread:3903645 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 05:52:14,697 INFO MainThread:3903645 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 05:52:14,699 INFO MainThread:3903645 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 05:52:14,705 INFO MainThread:3903645 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 05:52:45,224 INFO MainThread:3903645 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 05:56:43,091 INFO wandb-AsyncioManager-main:3903645 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 05:56:43,091 INFO wandb-AsyncioManager-main:3903645 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_055213-77omz43z/run-77omz43z.wandb b/wandb/run-20250914_055213-77omz43z/run-77omz43z.wandb new file mode 100644 index 0000000000000000000000000000000000000000..bdc87451a9785e1cd879152665b9502657f18f65 Binary files /dev/null and b/wandb/run-20250914_055213-77omz43z/run-77omz43z.wandb differ diff --git a/wandb/run-20250914_061839-0svspjiy/files/config.yaml b/wandb/run-20250914_061839-0svspjiy/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2dcbdd435577895a1b092602f62b3ba1b48865e0 --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + bkekzg58ihkgvs6dj0652alj3x9d7k8w: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.limit_val_batches=0 + - --trainer.val_check_interval=null + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3504419885056" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T22:18:39.317313Z" + writerId: bkekzg58ihkgvs6dj0652alj3x9d7k8w + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_061839-0svspjiy/files/output.log b/wandb/run-20250914_061839-0svspjiy/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c18923e8119e3fbeac207b237837c64a0a7b48aa --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/files/output.log @@ -0,0 +1,14 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 100%|███████████████████████████████████████████████████████████████████████████████████████████| 2/2 [01:55<00:00, 0.02it/s, v_num=pjiy] +`Trainer.fit` stopped: `max_epochs=1` reached. +✅ LoRA adapter saved to: checkpoints/archer_Qwen3-14B_rsa +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (5120) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. +Namespace(config=None, subcommand='fit', fit=Namespace(config=None, seed_everything=0, trainer=Namespace(accelerator='gpu', strategy='deepspeed_stage_3', devices='8', num_nodes=1, precision='bf16', logger=Namespace(class_path='lightning.pytorch.loggers.WandbLogger', init_args=Namespace(name='Test-AC-critic_expectile_0.9-inv_temp_1.0', save_dir='.', version=None, offline=False, dir=None, id=None, anonymous=None, project='RSAGame-Official', log_model=False, experiment=None, prefix='', checkpoint_name=None, entity=None, notes=None, tags=None, config=None, config_exclude_keys=None, config_include_keys=None, allow_val_change=None, group=None, job_type=None, force=None, reinit=None, resume=None, resume_from=None, fork_from=None, save_code=None, tensorboard=None, sync_tensorboard=None, monitor_gym=None, settings=None)), callbacks=None, fast_dev_run=False, max_epochs=1, min_epochs=None, max_steps=-1, min_steps=None, max_time=None, limit_train_batches=None, limit_val_batches=0, limit_test_batches=None, limit_predict_batches=None, overfit_batches=0.0, val_check_interval=None, check_val_every_n_epoch=1, num_sanity_val_steps=None, log_every_n_steps=None, enable_checkpointing=None, enable_progress_bar=None, enable_model_summary=False, accumulate_grad_batches=1, gradient_clip_val=None, gradient_clip_algorithm=None, deterministic=None, benchmark=None, inference_mode=True, use_distributed_sampler=True, profiler=None, detect_anomaly=False, barebones=False, plugins=None, sync_batchnorm=False, reload_dataloaders_every_n_epochs=0, default_root_dir='checkpoints/archer_Qwen3-14B_rsa', model_registry=None), model=Namespace(class_path='Algorithms.OfflineArcher', init_args=Namespace(model_name_or_path='/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', inv_temp=1.0, actor_lr=1e-05, critic_lr=1e-05, tau=0.05, accumulate_grad_batches=16, discount_factor=0.99, critic_expectile=0.9, optimize_critic=True, actor_checkpoint=None, critic_checkpoint=None)), data=Namespace(class_path='Tasks.RSAGame', init_args=Namespace(base_model='Qwen3-14B', batch_size=2, n_traj_eval=4, word_list=None)), optimizer=None, lr_scheduler=None, ckpt_path=None)) diff --git a/wandb/run-20250914_061839-0svspjiy/files/requirements.txt b/wandb/run-20250914_061839-0svspjiy/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_061839-0svspjiy/files/wandb-metadata.json b/wandb/run-20250914_061839-0svspjiy/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6bad64f81edaf33cf0d7a4ec9d31e92ff711b2f5 --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T22:18:39.317313Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3504419885056" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "bkekzg58ihkgvs6dj0652alj3x9d7k8w" +} \ No newline at end of file diff --git a/wandb/run-20250914_061839-0svspjiy/files/wandb-summary.json b/wandb/run-20250914_061839-0svspjiy/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..38b7e063e9be96dc7dfd769a8f43c754b070a4c9 --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":169},"_runtime":169} \ No newline at end of file diff --git a/wandb/run-20250914_061839-0svspjiy/logs/debug-core.log b/wandb/run-20250914_061839-0svspjiy/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..d37dc1fc676fa80c7f20f54bda28f8ca4f005c24 --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T06:18:39.374321976+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpxfvw71gm/port-3959047.txt","pid":3959047,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T06:18:39.374521778+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T06:18:39.375432055+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3959047} +{"time":"2025-09-14T06:18:39.375392708+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3959047-3971213-1651773266/socket","Net":"unix"}} +{"time":"2025-09-14T06:18:39.56362996+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T06:18:39.575479765+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"0svspjiy","id":"1(@)"} +{"time":"2025-09-14T06:18:40.065077802+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"0svspjiy","id":"1(@)"} +{"time":"2025-09-14T06:21:29.78247875+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T06:21:29.782634861+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T06:21:29.782756654+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T06:21:29.782802759+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T06:21:29.783479256+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3959047-3971213-1651773266/socket","Net":"unix"}} +{"time":"2025-09-14T06:21:31.342055558+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T06:21:31.342081851+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T06:21:31.34209322+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_061839-0svspjiy/logs/debug-internal.log b/wandb/run-20250914_061839-0svspjiy/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8408d3540422539a9d0056adca9abd6dca95bf0e --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T06:18:39.57574319+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T06:18:39.619842352+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T06:18:40.064990492+08:00","level":"INFO","msg":"stream: created new stream","id":"0svspjiy"} +{"time":"2025-09-14T06:18:40.065063068+08:00","level":"INFO","msg":"stream: started","id":"0svspjiy"} +{"time":"2025-09-14T06:18:40.065080972+08:00","level":"INFO","msg":"writer: started","stream_id":"0svspjiy"} +{"time":"2025-09-14T06:18:40.065268794+08:00","level":"INFO","msg":"handler: started","stream_id":"0svspjiy"} +{"time":"2025-09-14T06:18:40.065309705+08:00","level":"INFO","msg":"sender: started","stream_id":"0svspjiy"} +{"time":"2025-09-14T06:21:29.782635256+08:00","level":"INFO","msg":"stream: closing","id":"0svspjiy"} +{"time":"2025-09-14T06:21:31.081567874+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T06:21:31.340413066+08:00","level":"INFO","msg":"handler: closed","stream_id":"0svspjiy"} +{"time":"2025-09-14T06:21:31.341325059+08:00","level":"INFO","msg":"sender: closed","stream_id":"0svspjiy"} +{"time":"2025-09-14T06:21:31.341375281+08:00","level":"INFO","msg":"stream: closed","id":"0svspjiy"} diff --git a/wandb/run-20250914_061839-0svspjiy/logs/debug.log b/wandb/run-20250914_061839-0svspjiy/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..f26924222592fd6e6ef13ba9e622ac03626a5100 --- /dev/null +++ b/wandb/run-20250914_061839-0svspjiy/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 06:18:39,323 INFO MainThread:3959047 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 06:18:39,324 INFO MainThread:3959047 [wandb_setup.py:_flush():81] Configure stats pid to 3959047 +2025-09-14 06:18:39,324 INFO MainThread:3959047 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 06:18:39,324 INFO MainThread:3959047 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 06:18:39,324 INFO MainThread:3959047 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 06:18:39,324 INFO MainThread:3959047 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_061839-0svspjiy/logs/debug.log +2025-09-14 06:18:39,325 INFO MainThread:3959047 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_061839-0svspjiy/logs/debug-internal.log +2025-09-14 06:18:39,325 INFO MainThread:3959047 [wandb_init.py:init():813] calling init triggers +2025-09-14 06:18:39,325 INFO MainThread:3959047 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 06:18:39,325 INFO MainThread:3959047 [wandb_init.py:init():854] starting backend +2025-09-14 06:18:39,564 INFO MainThread:3959047 [wandb_init.py:init():857] sending inform_init request +2025-09-14 06:18:39,568 INFO MainThread:3959047 [wandb_init.py:init():865] backend started and connected +2025-09-14 06:18:39,572 INFO MainThread:3959047 [wandb_init.py:init():936] updated telemetry +2025-09-14 06:18:39,583 INFO MainThread:3959047 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 06:18:40,423 INFO MainThread:3959047 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 06:18:40,855 INFO MainThread:3959047 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 06:18:40,856 INFO MainThread:3959047 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 06:18:40,856 INFO MainThread:3959047 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 06:18:40,857 INFO MainThread:3959047 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 06:18:40,862 INFO MainThread:3959047 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 06:19:12,485 INFO MainThread:3959047 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 06:21:29,782 INFO wandb-AsyncioManager-main:3959047 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 06:21:29,783 INFO wandb-AsyncioManager-main:3959047 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_061839-0svspjiy/run-0svspjiy.wandb b/wandb/run-20250914_061839-0svspjiy/run-0svspjiy.wandb new file mode 100644 index 0000000000000000000000000000000000000000..6f41e4ad90b985e62eacab7ee85a88012f573080 Binary files /dev/null and b/wandb/run-20250914_061839-0svspjiy/run-0svspjiy.wandb differ diff --git a/wandb/run-20250914_063045-angc7phy/files/config.yaml b/wandb/run-20250914_063045-angc7phy/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f3c67efcd8df8519f188e24805d6e4697716b570 --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + a2vhuipwa547iowocz5uscghkwlgrrt0: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.limit_val_batches=0 + - --trainer.val_check_interval=null + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3535056084992" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T22:30:45.022494Z" + writerId: a2vhuipwa547iowocz5uscghkwlgrrt0 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_063045-angc7phy/files/output.log b/wandb/run-20250914_063045-angc7phy/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..97291d83dbf3c6c4291084b062d63ca667afdfa5 --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/files/output.log @@ -0,0 +1,14 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 100%|██████████████████████████████| 2/2 [02:00<00:00, 0.02it/s, v_num=7phy] +`Trainer.fit` stopped: `max_epochs=1` reached. +✅ LoRA adapter saved to: checkpoints/archer_Qwen3-14B_rsa +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (5120) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. diff --git a/wandb/run-20250914_063045-angc7phy/files/requirements.txt b/wandb/run-20250914_063045-angc7phy/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..9fed1510572a979e84b9a3683106b3c8c1e3126e --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/files/requirements.txt @@ -0,0 +1,139 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +archer==0.1.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_063045-angc7phy/files/wandb-metadata.json b/wandb/run-20250914_063045-angc7phy/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..1fc90233bc800bd87cb6ef562ab06e658aeeca0d --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T22:30:45.022494Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3535056084992" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "a2vhuipwa547iowocz5uscghkwlgrrt0" +} \ No newline at end of file diff --git a/wandb/run-20250914_063045-angc7phy/files/wandb-summary.json b/wandb/run-20250914_063045-angc7phy/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e1721ce9e839afb658f3a2ff42726d78d7f9c78e --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":163,"_wandb":{"runtime":163}} \ No newline at end of file diff --git a/wandb/run-20250914_063045-angc7phy/logs/debug-core.log b/wandb/run-20250914_063045-angc7phy/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..f79bcb7be7b966ec19040392aea5e38239d5153d --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T06:30:45.076623565+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmptipm4gwr/port-3990680.txt","pid":3990680,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T06:30:45.076829796+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T06:30:45.077514577+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3990680} +{"time":"2025-09-14T06:30:45.077500579+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3990680-4003994-932785811/socket","Net":"unix"}} +{"time":"2025-09-14T06:30:45.266326256+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T06:30:45.297428535+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"angc7phy","id":"1(@)"} +{"time":"2025-09-14T06:30:45.794954496+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"angc7phy","id":"1(@)"} +{"time":"2025-09-14T06:33:29.752604331+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T06:33:29.752768857+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T06:33:29.752816772+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T06:33:29.752954916+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T06:33:29.753478386+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3990680-4003994-932785811/socket","Net":"unix"}} +{"time":"2025-09-14T06:33:31.340987612+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T06:33:31.34104873+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T06:33:31.341068946+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_063045-angc7phy/logs/debug-internal.log b/wandb/run-20250914_063045-angc7phy/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..f10afadc2fbdcf3b7b7c8f1d0098c61cb451ebc3 --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T06:30:45.297807457+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T06:30:45.339580375+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T06:30:45.794849639+08:00","level":"INFO","msg":"stream: created new stream","id":"angc7phy"} +{"time":"2025-09-14T06:30:45.7949399+08:00","level":"INFO","msg":"stream: started","id":"angc7phy"} +{"time":"2025-09-14T06:30:45.794974931+08:00","level":"INFO","msg":"writer: started","stream_id":"angc7phy"} +{"time":"2025-09-14T06:30:45.79519463+08:00","level":"INFO","msg":"handler: started","stream_id":"angc7phy"} +{"time":"2025-09-14T06:30:45.79519917+08:00","level":"INFO","msg":"sender: started","stream_id":"angc7phy"} +{"time":"2025-09-14T06:33:29.752697483+08:00","level":"INFO","msg":"stream: closing","id":"angc7phy"} +{"time":"2025-09-14T06:33:31.086802302+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T06:33:31.338796501+08:00","level":"INFO","msg":"handler: closed","stream_id":"angc7phy"} +{"time":"2025-09-14T06:33:31.339689931+08:00","level":"INFO","msg":"sender: closed","stream_id":"angc7phy"} +{"time":"2025-09-14T06:33:31.339758993+08:00","level":"INFO","msg":"stream: closed","id":"angc7phy"} diff --git a/wandb/run-20250914_063045-angc7phy/logs/debug.log b/wandb/run-20250914_063045-angc7phy/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..fedefc521f2fec4d262795605ce8a3bedba7ff60 --- /dev/null +++ b/wandb/run-20250914_063045-angc7phy/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 06:30:45,027 INFO MainThread:3990680 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 06:30:45,028 INFO MainThread:3990680 [wandb_setup.py:_flush():81] Configure stats pid to 3990680 +2025-09-14 06:30:45,028 INFO MainThread:3990680 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 06:30:45,028 INFO MainThread:3990680 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 06:30:45,028 INFO MainThread:3990680 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 06:30:45,028 INFO MainThread:3990680 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_063045-angc7phy/logs/debug.log +2025-09-14 06:30:45,029 INFO MainThread:3990680 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_063045-angc7phy/logs/debug-internal.log +2025-09-14 06:30:45,029 INFO MainThread:3990680 [wandb_init.py:init():813] calling init triggers +2025-09-14 06:30:45,029 INFO MainThread:3990680 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 06:30:45,029 INFO MainThread:3990680 [wandb_init.py:init():854] starting backend +2025-09-14 06:30:45,267 INFO MainThread:3990680 [wandb_init.py:init():857] sending inform_init request +2025-09-14 06:30:45,279 INFO MainThread:3990680 [wandb_init.py:init():865] backend started and connected +2025-09-14 06:30:45,293 INFO MainThread:3990680 [wandb_init.py:init():936] updated telemetry +2025-09-14 06:30:45,323 INFO MainThread:3990680 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 06:30:46,145 INFO MainThread:3990680 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 06:30:46,586 INFO MainThread:3990680 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 06:30:46,587 INFO MainThread:3990680 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 06:30:46,587 INFO MainThread:3990680 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 06:30:46,587 INFO MainThread:3990680 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 06:30:46,594 INFO MainThread:3990680 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 06:31:20,240 INFO MainThread:3990680 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 06:33:29,753 INFO wandb-AsyncioManager-main:3990680 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 06:33:29,753 INFO wandb-AsyncioManager-main:3990680 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_063045-angc7phy/run-angc7phy.wandb b/wandb/run-20250914_063045-angc7phy/run-angc7phy.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e27a2d0435b2d4faca3a96b1a5db05f54b9e8907 Binary files /dev/null and b/wandb/run-20250914_063045-angc7phy/run-angc7phy.wandb differ diff --git a/wandb/run-20250914_064024-eidx0owo/files/config.yaml b/wandb/run-20250914_064024-eidx0owo/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..11f0c93955e7395bd6962776ee5f992fbf83aeaf --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + vgdhabchx3m4ahtakn9t3u6xsalc3uts: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.n_traj_eval=4 + - --data.base_model=Qwen3-14B + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.limit_val_batches=0 + - --trainer.val_check_interval=null + - --trainer.enable_model_summary=false + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3565639372800" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T22:40:24.526016Z" + writerId: vgdhabchx3m4ahtakn9t3u6xsalc3uts + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_064024-eidx0owo/files/output.log b/wandb/run-20250914_064024-eidx0owo/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..f3aaa722f09cae52a35055083e43c02b772fcbf6 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/files/output.log @@ -0,0 +1,240 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (2) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/2 [00:00 + return loss, log + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 597, in awr_loss + return super().configure_optimizers() + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 100, in get_logsum_prob + generated_probabilities = self.to_tokens_and_logprobs(alltext) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 132, in to_tokens_and_logprobs + outputs = self.model(input_ids) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward + return self.base_model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward + return self.model.forward(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 940, in wrapper + output = func(self, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward + outputs: BaseModelOutputWithPast = self.model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 1064, in wrapper + outputs = func(self, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 410, in forward + hidden_states = decoder_layer( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ + return super().__call__(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 275, in forward + hidden_states = self.mlp(hidden_states) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 82, in forward + down_proj = self.down_proj(self.act_fn(self.gate_proj(x)) * self.up_proj(x)) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1816, in inner + args_result = hook(self, args) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 929, in _fn + return fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/parameter_offload.py", line 298, in _pre_forward_module_hook + self.pre_sub_module_forward_function(module) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/parameter_offload.py", line 473, in pre_sub_module_forward_function + param_coordinator.fetch_sub_module(sub_module, forward=True) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_dynamo/eval_frame.py", line 929, in _fn + return fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/utils/_contextlib.py", line 120, in decorate_context + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partitioned_param_coordinator.py", line 317, in fetch_sub_module + self.__all_gather_params(params_to_fetch, forward) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partitioned_param_coordinator.py", line 475, in __all_gather_params + self.__all_gather_params_(nonquantized_params, forward, quantize=self.zero_quantized_weights) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partitioned_param_coordinator.py", line 504, in __all_gather_params_ + handle = param_group[0].all_gather_coalesced(param_group, quantize=quantize) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partition_parameters.py", line 1332, in all_gather_coalesced + handles = _dist_allgather_fn( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/partition_parameters.py", line 109, in _dist_allgather_fn + return instrument_w_nvtx(dist.allgather_fn)(output_tensor, input_tensor, group=group, async_op=True) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/comm/comm.py", line 333, in allgather_fn + return all_gather_into_tensor(output_tensor, input_tensor, group=group, async_op=async_op, debug=debug) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/comm/comm.py", line 118, in log_wrapper + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/comm/comm.py", line 318, in all_gather_into_tensor + return cdb.all_gather_into_tensor(output_tensor=output_tensor, input_tensor=tensor, group=group, async_op=async_op) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/comm/torch.py", line 236, in all_gather_into_tensor + return self.all_gather_function(output_tensor=output_tensor, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/c10d_logger.py", line 81, in wrapper + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/distributed/distributed_c10d.py", line 3986, in all_gather_into_tensor + work = group._allgather_base(output_tensor, input_tensor, opts) +KeyboardInterrupt + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 501, in main + run() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 351, in run_file + runpy.run_path(target, run_name="__main__") + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 310, in run_path + return _run_module_code(code, init_globals, run_name, pkg_name=pkg_name, script_name=fname) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 127, in _run_module_code + _run_code(code, mod_globals, init_globals, mod_name, mod_spec, pkg_name, script_name) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/_vendored/pydevd/_pydevd_bundle/pydevd_runpy.py", line 118, in _run_code + exec(code, run_globals) + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 66, in _call_and_handle_interrupt + sys.exit(1) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/wandb/sdk/lib/exit_hooks.py", line 36, in exit + self._orig_exit(orig_code) # type: ignore +SystemExit: 1 + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/__main__.py", line 71, in + cli.main() + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/server/cli.py", line 503, in main + log.reraise_exception( + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/common/log.py", line 222, in reraise_exception + _exception(format_string, *args, **kwargs) + File "/home/jiashuo/.cursor-server/extensions/ms-python.debugpy-2025.10.0-linux-x64/bundled/libs/debugpy/adapter/../../debugpy/launcher/../../debugpy/../debugpy/common/log.py", line 188, in _exception + exception = "".join(traceback.format_exception(*exc_info)) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/traceback.py", line 135, in format_exception + te = TracebackException(type(value), value, tb, limit=limit, compact=True) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/exceptiongroup/_formatting.py", line 179, in __init__ + context = PatchedTracebackException( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/exceptiongroup/_formatting.py", line 96, in __init__ + self.stack = traceback.StackSummary.extract( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/traceback.py", line 383, in extract + f.line + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/traceback.py", line 306, in line + self._line = linecache.getline(self.filename, self.lineno) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/linecache.py", line 30, in getline + lines = getlines(filename, module_globals) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/linecache.py", line 46, in getlines + return updatecache(filename, module_globals) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/linecache.py", line 136, in updatecache + with tokenize.open(fullname) as fp: +KeyboardInterrupt +Exception ignored in: diff --git a/wandb/run-20250914_064024-eidx0owo/files/requirements.txt b/wandb/run-20250914_064024-eidx0owo/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..dcafaeb778c74404b74331eb650c018a7e2a56d9 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/files/requirements.txt @@ -0,0 +1,140 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_064024-eidx0owo/files/wandb-metadata.json b/wandb/run-20250914_064024-eidx0owo/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8f327240383794b119707d49d97010ed63764a12 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T22:40:24.526016Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3565639372800" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "vgdhabchx3m4ahtakn9t3u6xsalc3uts" +} \ No newline at end of file diff --git a/wandb/run-20250914_064024-eidx0owo/files/wandb-summary.json b/wandb/run-20250914_064024-eidx0owo/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..7b7f6bbcb78630a18aad85b8e448b0042f42c288 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":39},"_runtime":39} \ No newline at end of file diff --git a/wandb/run-20250914_064024-eidx0owo/logs/debug-core.log b/wandb/run-20250914_064024-eidx0owo/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..4548b0998d33f71d4f505b5cebf4b005686b6b63 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-14T06:40:24.614108971+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpd2i2z2ia/port-4024003.txt","pid":4024003,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T06:40:24.614655327+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T06:40:24.61592934+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4024003-4035677-1490663784/socket","Net":"unix"}} +{"time":"2025-09-14T06:40:24.616130731+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4024003} +{"time":"2025-09-14T06:40:24.79841284+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T06:40:24.818397006+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"eidx0owo","id":"1(@)"} +{"time":"2025-09-14T06:40:25.319662804+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"eidx0owo","id":"1(@)"} +{"time":"2025-09-14T06:41:05.355685428+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T06:41:05.355780531+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T06:41:05.355821858+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T06:41:05.355883723+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T06:41:05.356484625+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4024003-4035677-1490663784/socket","Net":"unix"}} +{"time":"2025-09-14T06:41:06.248253483+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_064024-eidx0owo/logs/debug-internal.log b/wandb/run-20250914_064024-eidx0owo/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e8cccdb1fe6ae843c3c440da8a57a44ffbec2209 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T06:40:24.818637368+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T06:40:24.854275041+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T06:40:25.319503607+08:00","level":"INFO","msg":"stream: created new stream","id":"eidx0owo"} +{"time":"2025-09-14T06:40:25.319646984+08:00","level":"INFO","msg":"stream: started","id":"eidx0owo"} +{"time":"2025-09-14T06:40:25.319682219+08:00","level":"INFO","msg":"writer: started","stream_id":"eidx0owo"} +{"time":"2025-09-14T06:40:25.319814476+08:00","level":"INFO","msg":"sender: started","stream_id":"eidx0owo"} +{"time":"2025-09-14T06:40:25.319915751+08:00","level":"INFO","msg":"handler: started","stream_id":"eidx0owo"} +{"time":"2025-09-14T06:41:05.355796016+08:00","level":"INFO","msg":"stream: closing","id":"eidx0owo"} diff --git a/wandb/run-20250914_064024-eidx0owo/logs/debug.log b/wandb/run-20250914_064024-eidx0owo/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..963b95c6f6e9685dc3c5acb141f9d032cbcfe858 --- /dev/null +++ b/wandb/run-20250914_064024-eidx0owo/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 06:40:24,534 INFO MainThread:4024003 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 06:40:24,535 INFO MainThread:4024003 [wandb_setup.py:_flush():81] Configure stats pid to 4024003 +2025-09-14 06:40:24,536 INFO MainThread:4024003 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 06:40:24,536 INFO MainThread:4024003 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 06:40:24,536 INFO MainThread:4024003 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 06:40:24,537 INFO MainThread:4024003 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_064024-eidx0owo/logs/debug.log +2025-09-14 06:40:24,537 INFO MainThread:4024003 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_064024-eidx0owo/logs/debug-internal.log +2025-09-14 06:40:24,537 INFO MainThread:4024003 [wandb_init.py:init():813] calling init triggers +2025-09-14 06:40:24,538 INFO MainThread:4024003 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 06:40:24,538 INFO MainThread:4024003 [wandb_init.py:init():854] starting backend +2025-09-14 06:40:24,800 INFO MainThread:4024003 [wandb_init.py:init():857] sending inform_init request +2025-09-14 06:40:24,813 INFO MainThread:4024003 [wandb_init.py:init():865] backend started and connected +2025-09-14 06:40:24,829 INFO MainThread:4024003 [wandb_init.py:init():936] updated telemetry +2025-09-14 06:40:24,859 INFO MainThread:4024003 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 06:40:25,630 INFO MainThread:4024003 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 06:40:26,092 INFO MainThread:4024003 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 06:40:26,092 INFO MainThread:4024003 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 06:40:26,092 INFO MainThread:4024003 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 06:40:26,093 INFO MainThread:4024003 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 06:40:26,104 INFO MainThread:4024003 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 06:40:57,390 INFO MainThread:4024003 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 06:41:05,355 INFO wandb-AsyncioManager-main:4024003 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 06:41:05,356 INFO wandb-AsyncioManager-main:4024003 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_064024-eidx0owo/run-eidx0owo.wandb b/wandb/run-20250914_064024-eidx0owo/run-eidx0owo.wandb new file mode 100644 index 0000000000000000000000000000000000000000..7b38f32178203dfcafac58c1ba68f645face2dc0 Binary files /dev/null and b/wandb/run-20250914_064024-eidx0owo/run-eidx0owo.wandb differ diff --git a/wandb/run-20250914_064732-huh8z1e4/files/config.yaml b/wandb/run-20250914_064732-huh8z1e4/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f3916f403c3c0d0b2e2d40868e8cdc5e3cfb3f9e --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + hg6vme9w9yc7kbzqbfyne99uov26ngdy: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3565640847360" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T22:47:32.593757Z" + writerId: hg6vme9w9yc7kbzqbfyne99uov26ngdy + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_064732-huh8z1e4/files/output.log b/wandb/run-20250914_064732-huh8z1e4/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ccaa8ea9966eeee1041042b2c8aef777044b949d --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/files/output.log @@ -0,0 +1,12 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 1%| | 45/5122 [19:11<36:04:42, 0.04it/s, v_num=z1e4] + +Detected KeyboardInterrupt, attempting graceful shutdown ... +[rank: 0] Received SIGTERM: 15 diff --git a/wandb/run-20250914_064732-huh8z1e4/files/requirements.txt b/wandb/run-20250914_064732-huh8z1e4/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_064732-huh8z1e4/files/wandb-metadata.json b/wandb/run-20250914_064732-huh8z1e4/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8438c21bfb9602d5511a145a79c211072d4d5ba4 --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T22:47:32.593757Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3565640847360" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "hg6vme9w9yc7kbzqbfyne99uov26ngdy" +} \ No newline at end of file diff --git a/wandb/run-20250914_064732-huh8z1e4/files/wandb-summary.json b/wandb/run-20250914_064732-huh8z1e4/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..2d145deec50c5799f22055d9b444765dd2e5edcd --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":1194},"_runtime":1194} \ No newline at end of file diff --git a/wandb/run-20250914_064732-huh8z1e4/logs/debug-core.log b/wandb/run-20250914_064732-huh8z1e4/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..7a917c04937d2b866e4de7065249d4b91a4cc6b0 --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T06:47:32.628985061+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp02llw5ls/port-4051194.txt","pid":4051194,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T06:47:32.629279668+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T06:47:32.630434774+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4051194} +{"time":"2025-09-14T06:47:32.630452668+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4051194-4052093-1920293626/socket","Net":"unix"}} +{"time":"2025-09-14T06:47:32.81858201+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T06:47:32.826823782+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"huh8z1e4","id":"1(@)"} +{"time":"2025-09-14T06:47:33.297725582+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"huh8z1e4","id":"1(@)"} +{"time":"2025-09-14T07:07:28.217301322+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T07:07:28.217475162+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T07:07:28.217624418+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T07:07:28.217615626+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T07:07:28.217986964+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4051194-4052093-1920293626/socket","Net":"unix"}} +{"time":"2025-09-14T07:07:29.750553315+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T07:07:29.750602712+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T07:07:29.750623432+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_064732-huh8z1e4/logs/debug-internal.log b/wandb/run-20250914_064732-huh8z1e4/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..29e8b453ee78135812c138f80ac4c5c92818043d --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T06:47:32.827009261+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T06:47:32.856308695+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T06:47:33.297627258+08:00","level":"INFO","msg":"stream: created new stream","id":"huh8z1e4"} +{"time":"2025-09-14T06:47:33.297712982+08:00","level":"INFO","msg":"stream: started","id":"huh8z1e4"} +{"time":"2025-09-14T06:47:33.297738961+08:00","level":"INFO","msg":"handler: started","stream_id":"huh8z1e4"} +{"time":"2025-09-14T06:47:33.297739367+08:00","level":"INFO","msg":"writer: started","stream_id":"huh8z1e4"} +{"time":"2025-09-14T06:47:33.297784235+08:00","level":"INFO","msg":"sender: started","stream_id":"huh8z1e4"} +{"time":"2025-09-14T07:07:28.217425142+08:00","level":"INFO","msg":"stream: closing","id":"huh8z1e4"} +{"time":"2025-09-14T07:07:29.526957321+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T07:07:29.748732361+08:00","level":"INFO","msg":"handler: closed","stream_id":"huh8z1e4"} +{"time":"2025-09-14T07:07:29.749761279+08:00","level":"INFO","msg":"sender: closed","stream_id":"huh8z1e4"} +{"time":"2025-09-14T07:07:29.749812478+08:00","level":"INFO","msg":"stream: closed","id":"huh8z1e4"} diff --git a/wandb/run-20250914_064732-huh8z1e4/logs/debug.log b/wandb/run-20250914_064732-huh8z1e4/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4f1118c0d2ae08f85d66679abbdde0a58feddd41 --- /dev/null +++ b/wandb/run-20250914_064732-huh8z1e4/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 06:47:32,597 INFO MainThread:4051194 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_setup.py:_flush():81] Configure stats pid to 4051194 +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_064732-huh8z1e4/logs/debug.log +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_064732-huh8z1e4/logs/debug-internal.log +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_init.py:init():813] calling init triggers +2025-09-14 06:47:32,598 INFO MainThread:4051194 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 06:47:32,599 INFO MainThread:4051194 [wandb_init.py:init():854] starting backend +2025-09-14 06:47:32,818 INFO MainThread:4051194 [wandb_init.py:init():857] sending inform_init request +2025-09-14 06:47:32,823 INFO MainThread:4051194 [wandb_init.py:init():865] backend started and connected +2025-09-14 06:47:32,825 INFO MainThread:4051194 [wandb_init.py:init():936] updated telemetry +2025-09-14 06:47:32,836 INFO MainThread:4051194 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 06:47:33,565 INFO MainThread:4051194 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 06:47:33,731 INFO MainThread:4051194 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 06:47:33,731 INFO MainThread:4051194 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 06:47:33,732 INFO MainThread:4051194 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 06:47:33,732 INFO MainThread:4051194 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 06:47:33,735 INFO MainThread:4051194 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 06:48:04,625 INFO MainThread:4051194 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 07:07:28,217 INFO wandb-AsyncioManager-main:4051194 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 07:07:28,217 INFO wandb-AsyncioManager-main:4051194 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_064732-huh8z1e4/run-huh8z1e4.wandb b/wandb/run-20250914_064732-huh8z1e4/run-huh8z1e4.wandb new file mode 100644 index 0000000000000000000000000000000000000000..a1532741f8ee0a8115d82600e279c454c364cc14 Binary files /dev/null and b/wandb/run-20250914_064732-huh8z1e4/run-huh8z1e4.wandb differ diff --git a/wandb/run-20250914_070804-9jn02hsn/files/output.log b/wandb/run-20250914_070804-9jn02hsn/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7b019a4572f3957b4e1359cfd85b7e0cd2941878 --- /dev/null +++ b/wandb/run-20250914_070804-9jn02hsn/files/output.log @@ -0,0 +1,10 @@ +19552 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/2561 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 129, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 151, in read_data + for game in random.sample(data, 1000): +NameError: name 'random' is not defined +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 129, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 151, in read_data +[rank0]: for game in random.sample(data, 1000): +[rank0]: NameError: name 'random' is not defined diff --git a/wandb/run-20250914_071257-09e1i0ge/files/requirements.txt b/wandb/run-20250914_071257-09e1i0ge/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_071257-09e1i0ge/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_071257-09e1i0ge/files/wandb-metadata.json b/wandb/run-20250914_071257-09e1i0ge/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8dba385d39465313c50b12bfac9913b8f194a0aa --- /dev/null +++ b/wandb/run-20250914_071257-09e1i0ge/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T23:12:57.342419Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=4", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3565639942144" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "fhys66qes9o5gh71f7lup4jykerdcebx" +} \ No newline at end of file diff --git a/wandb/run-20250914_071257-09e1i0ge/files/wandb-summary.json b/wandb/run-20250914_071257-09e1i0ge/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250914_071257-09e1i0ge/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250914_071257-09e1i0ge/logs/debug-core.log b/wandb/run-20250914_071257-09e1i0ge/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..37644f25e8f4e6849618090fb5c8322459e5944d --- /dev/null +++ b/wandb/run-20250914_071257-09e1i0ge/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T07:12:57.378151925+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmphyuiav6x/port-4089264.txt","pid":4089264,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T07:12:57.378411141+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T07:12:57.380925751+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4089264} +{"time":"2025-09-14T07:12:57.380986581+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4089264-4090006-2472838871/socket","Net":"unix"}} +{"time":"2025-09-14T07:12:57.568004118+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T07:12:57.577022676+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"09e1i0ge","id":"1(@)"} +{"time":"2025-09-14T07:12:58.058900207+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"09e1i0ge","id":"1(@)"} +{"time":"2025-09-14T07:12:59.181957661+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T07:12:59.182170962+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T07:12:59.182159032+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T07:12:59.182388993+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T07:12:59.182540034+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4089264-4090006-2472838871/socket","Net":"unix"}} +{"time":"2025-09-14T07:13:00.771315485+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T07:13:00.771367631+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T07:13:00.771433966+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_071257-09e1i0ge/logs/debug-internal.log b/wandb/run-20250914_071257-09e1i0ge/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..9dbcc7863f395e55cc79f53ddcc71741547c2868 --- /dev/null +++ b/wandb/run-20250914_071257-09e1i0ge/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T07:12:57.577182619+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T07:12:57.60076432+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T07:12:58.058777122+08:00","level":"INFO","msg":"stream: created new stream","id":"09e1i0ge"} +{"time":"2025-09-14T07:12:58.0588832+08:00","level":"INFO","msg":"stream: started","id":"09e1i0ge"} +{"time":"2025-09-14T07:12:58.05892956+08:00","level":"INFO","msg":"handler: started","stream_id":"09e1i0ge"} +{"time":"2025-09-14T07:12:58.058918497+08:00","level":"INFO","msg":"writer: started","stream_id":"09e1i0ge"} +{"time":"2025-09-14T07:12:58.05901693+08:00","level":"INFO","msg":"sender: started","stream_id":"09e1i0ge"} +{"time":"2025-09-14T07:12:59.182045121+08:00","level":"INFO","msg":"stream: closing","id":"09e1i0ge"} +{"time":"2025-09-14T07:13:00.519815285+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T07:13:00.769028809+08:00","level":"INFO","msg":"handler: closed","stream_id":"09e1i0ge"} +{"time":"2025-09-14T07:13:00.769577882+08:00","level":"INFO","msg":"sender: closed","stream_id":"09e1i0ge"} +{"time":"2025-09-14T07:13:00.769649155+08:00","level":"INFO","msg":"stream: closed","id":"09e1i0ge"} diff --git a/wandb/run-20250914_071257-09e1i0ge/logs/debug.log b/wandb/run-20250914_071257-09e1i0ge/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..f35d4cc85be4988cb2d1dd87353ccec5e320e7bc --- /dev/null +++ b/wandb/run-20250914_071257-09e1i0ge/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 07:12:57,346 INFO MainThread:4089264 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_setup.py:_flush():81] Configure stats pid to 4089264 +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_071257-09e1i0ge/logs/debug.log +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_071257-09e1i0ge/logs/debug-internal.log +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_init.py:init():813] calling init triggers +2025-09-14 07:12:57,347 INFO MainThread:4089264 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 07:12:57,348 INFO MainThread:4089264 [wandb_init.py:init():854] starting backend +2025-09-14 07:12:57,568 INFO MainThread:4089264 [wandb_init.py:init():857] sending inform_init request +2025-09-14 07:12:57,573 INFO MainThread:4089264 [wandb_init.py:init():865] backend started and connected +2025-09-14 07:12:57,575 INFO MainThread:4089264 [wandb_init.py:init():936] updated telemetry +2025-09-14 07:12:57,587 INFO MainThread:4089264 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 07:12:58,341 INFO MainThread:4089264 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 07:12:58,508 INFO MainThread:4089264 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 07:12:58,508 INFO MainThread:4089264 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 07:12:58,509 INFO MainThread:4089264 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 07:12:58,509 INFO MainThread:4089264 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 07:12:58,512 INFO MainThread:4089264 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 07:12:59,182 INFO wandb-AsyncioManager-main:4089264 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 07:12:59,182 INFO wandb-AsyncioManager-main:4089264 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_071257-09e1i0ge/run-09e1i0ge.wandb b/wandb/run-20250914_071257-09e1i0ge/run-09e1i0ge.wandb new file mode 100644 index 0000000000000000000000000000000000000000..909f8ca78807b15576190c13334c2bf342a554f2 Binary files /dev/null and b/wandb/run-20250914_071257-09e1i0ge/run-09e1i0ge.wandb differ diff --git a/wandb/run-20250914_071323-l12uaqab/files/config.yaml b/wandb/run-20250914_071323-l12uaqab/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b908d02158a571fa7a2b08f2a58e15217b01cd2a --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + pcpwkmx9ik4l4cw85mik7zu1zv2hng4o: + args: + - fit + - --data=RSAGame + - --data.batch_size=4 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3565640056832" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T23:13:23.478960Z" + writerId: pcpwkmx9ik4l4cw85mik7zu1zv2hng4o + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_071323-l12uaqab/files/output.log b/wandb/run-20250914_071323-l12uaqab/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1f04bcd250fa12938355f04c8486cea031cc39cf --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/files/output.log @@ -0,0 +1,57 @@ +19561 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 129, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 151, in read_data + for game in random.sample(data, 1000): +NameError: name 'random' is not defined +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 129, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 151, in read_data +[rank0]: for game in random.sample(data, 1000): +[rank0]: NameError: name 'random' is not defined diff --git a/wandb/run-20250914_071323-l12uaqab/files/requirements.txt b/wandb/run-20250914_071323-l12uaqab/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_071323-l12uaqab/files/wandb-metadata.json b/wandb/run-20250914_071323-l12uaqab/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a12aafaa74b55675f31524cfabd634e273d8da58 --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T23:13:23.478960Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=4", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3565640056832" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "pcpwkmx9ik4l4cw85mik7zu1zv2hng4o" +} \ No newline at end of file diff --git a/wandb/run-20250914_071323-l12uaqab/files/wandb-summary.json b/wandb/run-20250914_071323-l12uaqab/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250914_071323-l12uaqab/logs/debug-core.log b/wandb/run-20250914_071323-l12uaqab/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..8f0146a9c36f5b2cbc94f42a58ff9a8bbc04fee7 --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T07:13:23.519796102+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp9gi3t12m/port-4090952.txt","pid":4090952,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T07:13:23.519997032+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T07:13:23.520827911+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4090952} +{"time":"2025-09-14T07:13:23.520820061+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4090952-4091807-720925536/socket","Net":"unix"}} +{"time":"2025-09-14T07:13:23.706551795+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T07:13:23.714771171+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"l12uaqab","id":"1(@)"} +{"time":"2025-09-14T07:13:24.195285654+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"l12uaqab","id":"1(@)"} +{"time":"2025-09-14T07:13:25.213381061+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T07:13:25.213466298+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T07:13:25.213459136+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T07:13:25.213550546+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T07:13:25.213725012+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4090952-4091807-720925536/socket","Net":"unix"}} +{"time":"2025-09-14T07:13:26.73459458+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T07:13:26.734627436+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T07:13:26.734659253+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_071323-l12uaqab/logs/debug-internal.log b/wandb/run-20250914_071323-l12uaqab/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..fef68dfc67aabd4b27bc438aced1d269b4f8593d --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T07:13:23.714922428+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T07:13:23.738213222+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T07:13:24.195172452+08:00","level":"INFO","msg":"stream: created new stream","id":"l12uaqab"} +{"time":"2025-09-14T07:13:24.195270892+08:00","level":"INFO","msg":"stream: started","id":"l12uaqab"} +{"time":"2025-09-14T07:13:24.195314+08:00","level":"INFO","msg":"writer: started","stream_id":"l12uaqab"} +{"time":"2025-09-14T07:13:24.195360404+08:00","level":"INFO","msg":"handler: started","stream_id":"l12uaqab"} +{"time":"2025-09-14T07:13:24.195322193+08:00","level":"INFO","msg":"sender: started","stream_id":"l12uaqab"} +{"time":"2025-09-14T07:13:25.213472981+08:00","level":"INFO","msg":"stream: closing","id":"l12uaqab"} +{"time":"2025-09-14T07:13:26.484934093+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T07:13:26.732361651+08:00","level":"INFO","msg":"handler: closed","stream_id":"l12uaqab"} +{"time":"2025-09-14T07:13:26.73299581+08:00","level":"INFO","msg":"sender: closed","stream_id":"l12uaqab"} +{"time":"2025-09-14T07:13:26.733019281+08:00","level":"INFO","msg":"stream: closed","id":"l12uaqab"} diff --git a/wandb/run-20250914_071323-l12uaqab/logs/debug.log b/wandb/run-20250914_071323-l12uaqab/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..d64126db97de1dce490177918858ef2ac4931a7e --- /dev/null +++ b/wandb/run-20250914_071323-l12uaqab/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_setup.py:_flush():81] Configure stats pid to 4090952 +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_071323-l12uaqab/logs/debug.log +2025-09-14 07:13:23,484 INFO MainThread:4090952 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_071323-l12uaqab/logs/debug-internal.log +2025-09-14 07:13:23,485 INFO MainThread:4090952 [wandb_init.py:init():813] calling init triggers +2025-09-14 07:13:23,485 INFO MainThread:4090952 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 07:13:23,485 INFO MainThread:4090952 [wandb_init.py:init():854] starting backend +2025-09-14 07:13:23,707 INFO MainThread:4090952 [wandb_init.py:init():857] sending inform_init request +2025-09-14 07:13:23,712 INFO MainThread:4090952 [wandb_init.py:init():865] backend started and connected +2025-09-14 07:13:23,716 INFO MainThread:4090952 [wandb_init.py:init():936] updated telemetry +2025-09-14 07:13:23,728 INFO MainThread:4090952 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 07:13:24,480 INFO MainThread:4090952 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 07:13:24,594 INFO MainThread:4090952 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 07:13:24,594 INFO MainThread:4090952 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 07:13:24,594 INFO MainThread:4090952 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 07:13:24,594 INFO MainThread:4090952 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 07:13:24,597 INFO MainThread:4090952 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 07:13:25,214 INFO wandb-AsyncioManager-main:4090952 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 07:13:25,214 INFO wandb-AsyncioManager-main:4090952 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_071323-l12uaqab/run-l12uaqab.wandb b/wandb/run-20250914_071323-l12uaqab/run-l12uaqab.wandb new file mode 100644 index 0000000000000000000000000000000000000000..914ed16d5fb949c94b929d9ba36ccccecbce3ac6 Binary files /dev/null and b/wandb/run-20250914_071323-l12uaqab/run-l12uaqab.wandb differ diff --git a/wandb/run-20250914_071419-eo66b86x/files/output.log b/wandb/run-20250914_071419-eo66b86x/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7db20c876c0e3a1009f99b9bdeb1f35f9bee5e73 --- /dev/null +++ b/wandb/run-20250914_071419-eo66b86x/files/output.log @@ -0,0 +1,10 @@ +19552 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 5%|▌ | 14/268 [06:03<1:50:00, 0.04it/s, v_num=b86x] diff --git a/wandb/run-20250914_071419-eo66b86x/files/requirements.txt b/wandb/run-20250914_071419-eo66b86x/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_071419-eo66b86x/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_071419-eo66b86x/files/wandb-metadata.json b/wandb/run-20250914_071419-eo66b86x/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7ef2987aae9ea278fd8514b5000819cfb67deeba --- /dev/null +++ b/wandb/run-20250914_071419-eo66b86x/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T23:14:19.172959Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3565640499200" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "i5b0iw4doahzdfa4vischw0pewqm30ip" +} \ No newline at end of file diff --git a/wandb/run-20250914_071419-eo66b86x/logs/debug-core.log b/wandb/run-20250914_071419-eo66b86x/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..67664d3261ace64716b5718b2d8ac56d93e98e73 --- /dev/null +++ b/wandb/run-20250914_071419-eo66b86x/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T07:14:19.208734505+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpr9nwaegn/port-4094141.txt","pid":4094141,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T07:14:19.208945067+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T07:14:19.209575402+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4094141} +{"time":"2025-09-14T07:14:19.209566918+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4094141-4094924-3164968849/socket","Net":"unix"}} +{"time":"2025-09-14T07:14:19.396629213+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T07:14:19.406870571+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"eo66b86x","id":"1(@)"} +{"time":"2025-09-14T07:14:19.891035597+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"eo66b86x","id":"1(@)"} +{"time":"2025-09-14T07:21:39.250138157+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_071419-eo66b86x/logs/debug-internal.log b/wandb/run-20250914_071419-eo66b86x/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..057578a8dcfaaac4761f422e2b1644b4837e34eb --- /dev/null +++ b/wandb/run-20250914_071419-eo66b86x/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T07:14:19.40723326+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T07:14:19.444714813+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T07:14:19.89091688+08:00","level":"INFO","msg":"stream: created new stream","id":"eo66b86x"} +{"time":"2025-09-14T07:14:19.891012178+08:00","level":"INFO","msg":"stream: started","id":"eo66b86x"} +{"time":"2025-09-14T07:14:19.891032441+08:00","level":"INFO","msg":"writer: started","stream_id":"eo66b86x"} +{"time":"2025-09-14T07:14:19.891075248+08:00","level":"INFO","msg":"handler: started","stream_id":"eo66b86x"} +{"time":"2025-09-14T07:14:19.891108629+08:00","level":"INFO","msg":"sender: started","stream_id":"eo66b86x"} diff --git a/wandb/run-20250914_071419-eo66b86x/logs/debug.log b/wandb/run-20250914_071419-eo66b86x/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c41c354942d19d17efccd69e1856ba1780996502 --- /dev/null +++ b/wandb/run-20250914_071419-eo66b86x/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-14 07:14:19,177 INFO MainThread:4094141 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 07:14:19,177 INFO MainThread:4094141 [wandb_setup.py:_flush():81] Configure stats pid to 4094141 +2025-09-14 07:14:19,177 INFO MainThread:4094141 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 07:14:19,177 INFO MainThread:4094141 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 07:14:19,177 INFO MainThread:4094141 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 07:14:19,177 INFO MainThread:4094141 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_071419-eo66b86x/logs/debug.log +2025-09-14 07:14:19,178 INFO MainThread:4094141 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_071419-eo66b86x/logs/debug-internal.log +2025-09-14 07:14:19,178 INFO MainThread:4094141 [wandb_init.py:init():813] calling init triggers +2025-09-14 07:14:19,178 INFO MainThread:4094141 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 07:14:19,178 INFO MainThread:4094141 [wandb_init.py:init():854] starting backend +2025-09-14 07:14:19,396 INFO MainThread:4094141 [wandb_init.py:init():857] sending inform_init request +2025-09-14 07:14:19,401 INFO MainThread:4094141 [wandb_init.py:init():865] backend started and connected +2025-09-14 07:14:19,404 INFO MainThread:4094141 [wandb_init.py:init():936] updated telemetry +2025-09-14 07:14:19,415 INFO MainThread:4094141 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 07:14:20,168 INFO MainThread:4094141 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 07:14:20,333 INFO MainThread:4094141 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 07:14:20,333 INFO MainThread:4094141 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 07:14:20,333 INFO MainThread:4094141 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 07:14:20,333 INFO MainThread:4094141 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 07:14:20,336 INFO MainThread:4094141 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 07:14:48,018 INFO MainThread:4094141 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250914_071419-eo66b86x/run-eo66b86x.wandb b/wandb/run-20250914_071419-eo66b86x/run-eo66b86x.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_072202-c40vakf9/files/config.yaml b/wandb/run-20250914_072202-c40vakf9/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..be9f078894c77b6d40130116006688c75fa1ca63 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + z7t5e87k05iyqfqu4sk6cexed8s3tztj: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3565641342976" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-13T23:22:02.659698Z" + writerId: z7t5e87k05iyqfqu4sk6cexed8s3tztj + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_072202-c40vakf9/files/output.log b/wandb/run-20250914_072202-c40vakf9/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..386f050810aff0b185f75170e93194cba296bb3e --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/files/output.log @@ -0,0 +1,23 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 473, numel = 3938312 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 879 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 100%|██████████| 258/258 [1:06:44<00:00, 0.06it/s, v_num=akf9]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_rsa +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 631, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. +`Trainer.fit` stopped: `max_epochs=1` reached. +Epoch 0: 100%|██████████| 258/258 [1:06:48<00:00, 0.06it/s, v_num=akf9] diff --git a/wandb/run-20250914_072202-c40vakf9/files/requirements.txt b/wandb/run-20250914_072202-c40vakf9/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_072202-c40vakf9/files/wandb-metadata.json b/wandb/run-20250914_072202-c40vakf9/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e5a75a790104f1bdb758635663af7f0454fc5372 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-13T23:22:02.659698Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3565641342976" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "z7t5e87k05iyqfqu4sk6cexed8s3tztj" +} \ No newline at end of file diff --git a/wandb/run-20250914_072202-c40vakf9/files/wandb-summary.json b/wandb/run-20250914_072202-c40vakf9/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b7c9bf4f6b077a5d527027ef877c0dfd1bc02f87 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/files/wandb-summary.json @@ -0,0 +1 @@ +{"critic/v1.min":-0.0006699562072753906,"critic/v2.min":-0.048309326171875,"critic/target_q2.max":-0.058380126953125,"critic/v2.mean":-0.043121337890625,"actor/factor.max":1.533203125,"_runtime":4026,"critic/v2.max":-0.0379180908203125,"critic/q2.mean":0.47509765625,"actor/loss":3.6357421875,"actor/log_prob.max":-17.529296875,"critic/q2.min":0.423828125,"critic/target_q2.min":-0.06317138671875,"critic/v2.loss":4.0884689951781183e-05,"critic/q1.min":0.2998046875,"critic/v1.loss":9.458065323997289e-05,"actor/advantages.mean":0.38330078125,"actor/advantages.max":0.425537109375,"critic/target_q1.mean":-0.0282440185546875,"actor/advantages.min":0.341064453125,"critic/q1.mean":0.33935546875,"critic/v1.max":0.005003929138183594,"critic/v1.mean":0.002162456512451172,"trainer/global_step":249,"critic/q2.max":0.52587890625,"_wandb":{"runtime":4026},"critic/q2.loss":0.16375732421875,"actor/factor.min":1.4072265625,"critic/q1.max":0.3798828125,"critic/target_q1.min":-0.029876708984375,"critic/target_q2.mean":-0.06072998046875,"_step":4,"_timestamp":1.7578096215983949e+09,"actor/log_prob.mean":-38.015625,"critic/target_q1.max":-0.026611328125,"critic/q1.loss":0.15728759765625,"actor/factor.mean":1.46875,"epoch":0,"actor/log_prob.min":-58.453125} \ No newline at end of file diff --git a/wandb/run-20250914_072202-c40vakf9/logs/debug-core.log b/wandb/run-20250914_072202-c40vakf9/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..84f6bd4dc3cef9791afcf427a35697c7fa4b18c1 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T07:22:02.685453433+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmphymi3jlj/port-4104872.txt","pid":4104872,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T07:22:02.685640196+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T07:22:02.686796647+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4104872} +{"time":"2025-09-14T07:22:02.686763205+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4104872-4105611-2404222954/socket","Net":"unix"}} +{"time":"2025-09-14T07:22:02.873323589+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T07:22:02.881180054+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"c40vakf9","id":"1(@)"} +{"time":"2025-09-14T07:22:03.396173123+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"c40vakf9","id":"1(@)"} +{"time":"2025-09-14T08:29:10.601667946+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T08:29:10.601766559+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T08:29:10.6017922+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T08:29:10.601916765+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T08:29:10.602491256+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4104872-4105611-2404222954/socket","Net":"unix"}} +{"time":"2025-09-14T08:29:12.194149302+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T08:29:12.194200274+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T08:29:12.194224307+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_072202-c40vakf9/logs/debug-internal.log b/wandb/run-20250914_072202-c40vakf9/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..5e65af450b678e1d35f059bff74654f69198e882 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T07:22:02.881490536+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T07:22:02.910695898+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T07:22:03.396050295+08:00","level":"INFO","msg":"stream: created new stream","id":"c40vakf9"} +{"time":"2025-09-14T07:22:03.396158443+08:00","level":"INFO","msg":"stream: started","id":"c40vakf9"} +{"time":"2025-09-14T07:22:03.396220548+08:00","level":"INFO","msg":"handler: started","stream_id":"c40vakf9"} +{"time":"2025-09-14T07:22:03.396250362+08:00","level":"INFO","msg":"sender: started","stream_id":"c40vakf9"} +{"time":"2025-09-14T07:22:03.396208163+08:00","level":"INFO","msg":"writer: started","stream_id":"c40vakf9"} +{"time":"2025-09-14T08:29:10.601732131+08:00","level":"INFO","msg":"stream: closing","id":"c40vakf9"} +{"time":"2025-09-14T08:29:11.950716404+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T08:29:12.192105314+08:00","level":"INFO","msg":"handler: closed","stream_id":"c40vakf9"} +{"time":"2025-09-14T08:29:12.192757153+08:00","level":"INFO","msg":"sender: closed","stream_id":"c40vakf9"} +{"time":"2025-09-14T08:29:12.19280706+08:00","level":"INFO","msg":"stream: closed","id":"c40vakf9"} diff --git a/wandb/run-20250914_072202-c40vakf9/logs/debug.log b/wandb/run-20250914_072202-c40vakf9/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..3fb019bd5ccf2a9debd92a705a81156f4bceab41 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 07:22:02,661 INFO MainThread:4104872 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 07:22:02,661 INFO MainThread:4104872 [wandb_setup.py:_flush():81] Configure stats pid to 4104872 +2025-09-14 07:22:02,661 INFO MainThread:4104872 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 07:22:02,661 INFO MainThread:4104872 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 07:22:02,661 INFO MainThread:4104872 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 07:22:02,662 INFO MainThread:4104872 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_072202-c40vakf9/logs/debug.log +2025-09-14 07:22:02,662 INFO MainThread:4104872 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_072202-c40vakf9/logs/debug-internal.log +2025-09-14 07:22:02,662 INFO MainThread:4104872 [wandb_init.py:init():813] calling init triggers +2025-09-14 07:22:02,662 INFO MainThread:4104872 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 07:22:02,662 INFO MainThread:4104872 [wandb_init.py:init():854] starting backend +2025-09-14 07:22:02,873 INFO MainThread:4104872 [wandb_init.py:init():857] sending inform_init request +2025-09-14 07:22:02,875 INFO MainThread:4104872 [wandb_init.py:init():865] backend started and connected +2025-09-14 07:22:02,877 INFO MainThread:4104872 [wandb_init.py:init():936] updated telemetry +2025-09-14 07:22:02,884 INFO MainThread:4104872 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 07:22:03,682 INFO MainThread:4104872 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 07:22:03,866 INFO MainThread:4104872 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 07:22:03,866 INFO MainThread:4104872 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 07:22:03,866 INFO MainThread:4104872 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 07:22:03,867 INFO MainThread:4104872 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 07:22:03,869 INFO MainThread:4104872 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 07:22:21,948 INFO MainThread:4104872 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 08:29:10,602 INFO wandb-AsyncioManager-main:4104872 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 08:29:10,602 INFO wandb-AsyncioManager-main:4104872 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_072202-c40vakf9/run-c40vakf9.wandb b/wandb/run-20250914_072202-c40vakf9/run-c40vakf9.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ed61e4f31326d8811c1407c9e76fe2f161f2e859 --- /dev/null +++ b/wandb/run-20250914_072202-c40vakf9/run-c40vakf9.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:984bad91c9312ef1bc1a60a86d8f810fd5684ff461d5ba4b56e68f8d735c6edf +size 216513 diff --git a/wandb/run-20250914_082937-gk8a8jfh/files/config.yaml b/wandb/run-20250914_082937-gk8a8jfh/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f3dd08d27feb8465045325e0b7413159a00f0dd2 --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/files/config.yaml @@ -0,0 +1,83 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + 2heu7db3mgog65d95da66kxjfrnygaad: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T00:29:37.475400Z" + writerId: 2heu7db3mgog65d95da66kxjfrnygaad + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_082937-gk8a8jfh/files/output.log b/wandb/run-20250914_082937-gk8a8jfh/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..5438f78539b8412dab5c7471a4503316afd44119 --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/files/output.log @@ -0,0 +1,60 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 230, in read_data + from word_taboo import get_game_outcome, randomly_convert_game_history_to_query + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 6, in + from textblob import TextBlob, Word +ModuleNotFoundError: No module named 'textblob' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 230, in read_data +[rank0]: from word_taboo import get_game_outcome, randomly_convert_game_history_to_query +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 6, in +[rank0]: from textblob import TextBlob, Word +[rank0]: ModuleNotFoundError: No module named 'textblob' diff --git a/wandb/run-20250914_082937-gk8a8jfh/files/requirements.txt b/wandb/run-20250914_082937-gk8a8jfh/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_082937-gk8a8jfh/files/wandb-metadata.json b/wandb/run-20250914_082937-gk8a8jfh/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..f7099dac76ae71805705ba1def2e4f17f722577c --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/files/wandb-metadata.json @@ -0,0 +1,47 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T00:29:37.475400Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "writerId": "2heu7db3mgog65d95da66kxjfrnygaad" +} \ No newline at end of file diff --git a/wandb/run-20250914_082937-gk8a8jfh/files/wandb-summary.json b/wandb/run-20250914_082937-gk8a8jfh/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..1d476fc88692f959c7a899096787abbc21a55dbc --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":0,"_wandb":{"runtime":0}} \ No newline at end of file diff --git a/wandb/run-20250914_082937-gk8a8jfh/logs/debug-core.log b/wandb/run-20250914_082937-gk8a8jfh/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..51f1940108e3c7f9d391afd64c969b6f3a90ed7e --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/logs/debug-core.log @@ -0,0 +1,16 @@ +{"time":"2025-09-14T08:29:37.514415428+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp73_bd6ad/port-4113498.txt","pid":4113498,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T08:29:37.514821818+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T08:29:37.515988876+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4113498-4113996-653341652/socket","Net":"unix"}} +{"time":"2025-09-14T08:29:37.516083363+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4113498} +{"time":"2025-09-14T08:29:37.703571525+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T08:29:37.716250881+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"gk8a8jfh","id":"1(@)"} +{"time":"2025-09-14T08:29:38.202699219+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"gk8a8jfh","id":"1(@)"} +{"time":"2025-09-14T08:29:38.741013486+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T08:29:38.74124529+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T08:29:38.741285386+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T08:29:38.741343184+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T08:29:38.742324266+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4113498-4113996-653341652/socket","Net":"unix"}} +{"time":"2025-09-14T08:29:38.941764333+08:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write unix /tmp/wandb-4113498-4113996-653341652/socket->@: use of closed network connection","id":"1(@)"} +{"time":"2025-09-14T08:29:40.629519751+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T08:29:40.629560375+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T08:29:40.62957724+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_082937-gk8a8jfh/logs/debug-internal.log b/wandb/run-20250914_082937-gk8a8jfh/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..6419d0287211dc4dd41da87324b262dedb6f8088 --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T08:29:37.71653723+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T08:29:37.744379078+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T08:29:38.202628023+08:00","level":"INFO","msg":"stream: created new stream","id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:38.202690645+08:00","level":"INFO","msg":"stream: started","id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:38.202832281+08:00","level":"INFO","msg":"writer: started","stream_id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:38.202890669+08:00","level":"INFO","msg":"sender: started","stream_id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:38.202922988+08:00","level":"INFO","msg":"handler: started","stream_id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:38.741068359+08:00","level":"INFO","msg":"stream: closing","id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:40.24802374+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T08:29:40.626822679+08:00","level":"INFO","msg":"handler: closed","stream_id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:40.627560503+08:00","level":"INFO","msg":"sender: closed","stream_id":"gk8a8jfh"} +{"time":"2025-09-14T08:29:40.627590412+08:00","level":"INFO","msg":"stream: closed","id":"gk8a8jfh"} diff --git a/wandb/run-20250914_082937-gk8a8jfh/logs/debug.log b/wandb/run-20250914_082937-gk8a8jfh/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..f938f30c302cc1541398bc6769207665a6ac5505 --- /dev/null +++ b/wandb/run-20250914_082937-gk8a8jfh/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 08:29:37,480 INFO MainThread:4113498 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 08:29:37,480 INFO MainThread:4113498 [wandb_setup.py:_flush():81] Configure stats pid to 4113498 +2025-09-14 08:29:37,480 INFO MainThread:4113498 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_082937-gk8a8jfh/logs/debug.log +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_082937-gk8a8jfh/logs/debug-internal.log +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_init.py:init():813] calling init triggers +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 08:29:37,481 INFO MainThread:4113498 [wandb_init.py:init():854] starting backend +2025-09-14 08:29:37,703 INFO MainThread:4113498 [wandb_init.py:init():857] sending inform_init request +2025-09-14 08:29:37,708 INFO MainThread:4113498 [wandb_init.py:init():865] backend started and connected +2025-09-14 08:29:37,711 INFO MainThread:4113498 [wandb_init.py:init():936] updated telemetry +2025-09-14 08:29:37,723 INFO MainThread:4113498 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 08:29:38,561 INFO MainThread:4113498 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 08:29:38,698 INFO MainThread:4113498 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 08:29:38,698 INFO MainThread:4113498 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 08:29:38,698 INFO MainThread:4113498 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 08:29:38,698 INFO MainThread:4113498 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 08:29:38,700 INFO MainThread:4113498 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 08:29:38,739 INFO wandb-AsyncioManager-main:4113498 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 08:29:38,740 INFO wandb-AsyncioManager-main:4113498 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250914_082937-gk8a8jfh/run-gk8a8jfh.wandb b/wandb/run-20250914_082937-gk8a8jfh/run-gk8a8jfh.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ee6b29f1889434536b6063ac11ba26137da5a40e Binary files /dev/null and b/wandb/run-20250914_082937-gk8a8jfh/run-gk8a8jfh.wandb differ diff --git a/wandb/run-20250914_083001-7x38f13a/files/config.yaml b/wandb/run-20250914_083001-7x38f13a/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d47dd479fa40cd1316065c96c935544a4eea2417 --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + ffwne1mtorujwf51ginl5xncev866o9y: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3584909574144" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T00:30:01.205330Z" + writerId: ffwne1mtorujwf51ginl5xncev866o9y + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_083001-7x38f13a/files/output.log b/wandb/run-20250914_083001-7x38f13a/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..5438f78539b8412dab5c7471a4503316afd44119 --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/files/output.log @@ -0,0 +1,60 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 230, in read_data + from word_taboo import get_game_outcome, randomly_convert_game_history_to_query + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 6, in + from textblob import TextBlob, Word +ModuleNotFoundError: No module named 'textblob' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 230, in read_data +[rank0]: from word_taboo import get_game_outcome, randomly_convert_game_history_to_query +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 6, in +[rank0]: from textblob import TextBlob, Word +[rank0]: ModuleNotFoundError: No module named 'textblob' diff --git a/wandb/run-20250914_083001-7x38f13a/files/requirements.txt b/wandb/run-20250914_083001-7x38f13a/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_083001-7x38f13a/files/wandb-metadata.json b/wandb/run-20250914_083001-7x38f13a/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..123ce56bd329b13d61a8b9cb5a9eb15e5b7331b1 --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T00:30:01.205330Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584909574144" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ffwne1mtorujwf51ginl5xncev866o9y" +} \ No newline at end of file diff --git a/wandb/run-20250914_083001-7x38f13a/files/wandb-summary.json b/wandb/run-20250914_083001-7x38f13a/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250914_083001-7x38f13a/logs/debug-core.log b/wandb/run-20250914_083001-7x38f13a/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..a1adead0085e73c239ff5df3b3f75ad2b7fe3e55 --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/logs/debug-core.log @@ -0,0 +1,16 @@ +{"time":"2025-09-14T08:30:01.240304055+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpg4hem77t/port-4114712.txt","pid":4114712,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T08:30:01.240512158+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T08:30:01.241295454+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4114712} +{"time":"2025-09-14T08:30:01.241308912+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4114712-4115205-3174814776/socket","Net":"unix"}} +{"time":"2025-09-14T08:30:01.428346589+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T08:30:01.436250256+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"7x38f13a","id":"1(@)"} +{"time":"2025-09-14T08:30:01.921662656+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"7x38f13a","id":"1(@)"} +{"time":"2025-09-14T08:30:02.434524191+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T08:30:02.43469768+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T08:30:02.434735792+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T08:30:02.434746939+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T08:30:02.435020228+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4114712-4115205-3174814776/socket","Net":"unix"}} +{"time":"2025-09-14T08:30:02.634658185+08:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write unix /tmp/wandb-4114712-4115205-3174814776/socket->@: use of closed network connection","id":"1(@)"} +{"time":"2025-09-14T08:30:04.313054548+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T08:30:04.31311073+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T08:30:04.313137698+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_083001-7x38f13a/logs/debug-internal.log b/wandb/run-20250914_083001-7x38f13a/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d4c226a7fa99b73587becbcdca07c3a1396d8bc4 --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T08:30:01.436453692+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T08:30:01.466018675+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T08:30:01.921534668+08:00","level":"INFO","msg":"stream: created new stream","id":"7x38f13a"} +{"time":"2025-09-14T08:30:01.921647256+08:00","level":"INFO","msg":"stream: started","id":"7x38f13a"} +{"time":"2025-09-14T08:30:01.921673416+08:00","level":"INFO","msg":"handler: started","stream_id":"7x38f13a"} +{"time":"2025-09-14T08:30:01.921714117+08:00","level":"INFO","msg":"writer: started","stream_id":"7x38f13a"} +{"time":"2025-09-14T08:30:01.921691477+08:00","level":"INFO","msg":"sender: started","stream_id":"7x38f13a"} +{"time":"2025-09-14T08:30:02.434584123+08:00","level":"INFO","msg":"stream: closing","id":"7x38f13a"} +{"time":"2025-09-14T08:30:04.02288829+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T08:30:04.310313473+08:00","level":"INFO","msg":"handler: closed","stream_id":"7x38f13a"} +{"time":"2025-09-14T08:30:04.311055513+08:00","level":"INFO","msg":"sender: closed","stream_id":"7x38f13a"} +{"time":"2025-09-14T08:30:04.311099671+08:00","level":"INFO","msg":"stream: closed","id":"7x38f13a"} diff --git a/wandb/run-20250914_083001-7x38f13a/logs/debug.log b/wandb/run-20250914_083001-7x38f13a/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7473206dc75c130a47a24227691afab737e4ad2f --- /dev/null +++ b/wandb/run-20250914_083001-7x38f13a/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 08:30:01,209 INFO MainThread:4114712 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 08:30:01,209 INFO MainThread:4114712 [wandb_setup.py:_flush():81] Configure stats pid to 4114712 +2025-09-14 08:30:01,209 INFO MainThread:4114712 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 08:30:01,209 INFO MainThread:4114712 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 08:30:01,209 INFO MainThread:4114712 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 08:30:01,210 INFO MainThread:4114712 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_083001-7x38f13a/logs/debug.log +2025-09-14 08:30:01,210 INFO MainThread:4114712 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_083001-7x38f13a/logs/debug-internal.log +2025-09-14 08:30:01,210 INFO MainThread:4114712 [wandb_init.py:init():813] calling init triggers +2025-09-14 08:30:01,210 INFO MainThread:4114712 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 08:30:01,210 INFO MainThread:4114712 [wandb_init.py:init():854] starting backend +2025-09-14 08:30:01,428 INFO MainThread:4114712 [wandb_init.py:init():857] sending inform_init request +2025-09-14 08:30:01,433 INFO MainThread:4114712 [wandb_init.py:init():865] backend started and connected +2025-09-14 08:30:01,436 INFO MainThread:4114712 [wandb_init.py:init():936] updated telemetry +2025-09-14 08:30:01,456 INFO MainThread:4114712 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 08:30:02,188 INFO MainThread:4114712 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 08:30:02,380 INFO MainThread:4114712 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 08:30:02,381 INFO MainThread:4114712 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 08:30:02,381 INFO MainThread:4114712 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 08:30:02,381 INFO MainThread:4114712 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 08:30:02,384 INFO MainThread:4114712 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 08:30:02,433 INFO wandb-AsyncioManager-main:4114712 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 08:30:02,433 INFO wandb-AsyncioManager-main:4114712 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250914_083001-7x38f13a/run-7x38f13a.wandb b/wandb/run-20250914_083001-7x38f13a/run-7x38f13a.wandb new file mode 100644 index 0000000000000000000000000000000000000000..618e274dfd4be336ccc2677f247906da236877a6 Binary files /dev/null and b/wandb/run-20250914_083001-7x38f13a/run-7x38f13a.wandb differ diff --git a/wandb/run-20250914_152953-h7gbi509/files/output.log b/wandb/run-20250914_152953-h7gbi509/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..89d356c2fa30cd4a50ed24759ae73eb1a21a3a3a --- /dev/null +++ b/wandb/run-20250914_152953-h7gbi509/files/output.log @@ -0,0 +1,9 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 5%|▌ | 14/268 [06:04<1:50:08, 0.04it/s, v_num=i509] diff --git a/wandb/run-20250914_152953-h7gbi509/files/requirements.txt b/wandb/run-20250914_152953-h7gbi509/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..70af9a8943aac0c1c2668b2ddd518e3280e95bff --- /dev/null +++ b/wandb/run-20250914_152953-h7gbi509/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_152953-h7gbi509/files/wandb-metadata.json b/wandb/run-20250914_152953-h7gbi509/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..65e8aac07ada918102e8d5ddd8efc2399a7844aa --- /dev/null +++ b/wandb/run-20250914_152953-h7gbi509/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T07:29:53.270078Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584823472128" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "8mnnv8t4gtc7xsn3sxzh5majyusvd0be" +} \ No newline at end of file diff --git a/wandb/run-20250914_152953-h7gbi509/logs/debug-core.log b/wandb/run-20250914_152953-h7gbi509/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..8f1048e6a61705b4e97e649c258b09551f1a5e49 --- /dev/null +++ b/wandb/run-20250914_152953-h7gbi509/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T15:29:53.307007247+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpu_9t27li/port-4148880.txt","pid":4148880,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T15:29:53.307222794+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T15:29:53.307764757+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4148880} +{"time":"2025-09-14T15:29:53.307772684+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4148880-4149663-749972919/socket","Net":"unix"}} +{"time":"2025-09-14T15:29:53.49468135+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T15:29:53.506417839+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"h7gbi509","id":"1(@)"} +{"time":"2025-09-14T15:29:53.994023777+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"h7gbi509","id":"1(@)"} +{"time":"2025-09-14T15:37:14.194886374+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_152953-h7gbi509/logs/debug-internal.log b/wandb/run-20250914_152953-h7gbi509/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0cdf46f2ff4ac731d0e048dbccf74595ab4e26fe --- /dev/null +++ b/wandb/run-20250914_152953-h7gbi509/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T15:29:53.506579476+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T15:29:53.537827743+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T15:29:53.99390818+08:00","level":"INFO","msg":"stream: created new stream","id":"h7gbi509"} +{"time":"2025-09-14T15:29:53.994006474+08:00","level":"INFO","msg":"stream: started","id":"h7gbi509"} +{"time":"2025-09-14T15:29:53.99403039+08:00","level":"INFO","msg":"writer: started","stream_id":"h7gbi509"} +{"time":"2025-09-14T15:29:53.99405692+08:00","level":"INFO","msg":"handler: started","stream_id":"h7gbi509"} +{"time":"2025-09-14T15:29:53.994090875+08:00","level":"INFO","msg":"sender: started","stream_id":"h7gbi509"} diff --git a/wandb/run-20250914_152953-h7gbi509/logs/debug.log b/wandb/run-20250914_152953-h7gbi509/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6d1821c86ee5c74a50a8cf266a4ffa6c1dfd960b --- /dev/null +++ b/wandb/run-20250914_152953-h7gbi509/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-14 15:29:53,274 INFO MainThread:4148880 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 15:29:53,274 INFO MainThread:4148880 [wandb_setup.py:_flush():81] Configure stats pid to 4148880 +2025-09-14 15:29:53,274 INFO MainThread:4148880 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 15:29:53,274 INFO MainThread:4148880 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 15:29:53,274 INFO MainThread:4148880 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 15:29:53,275 INFO MainThread:4148880 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_152953-h7gbi509/logs/debug.log +2025-09-14 15:29:53,275 INFO MainThread:4148880 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_152953-h7gbi509/logs/debug-internal.log +2025-09-14 15:29:53,275 INFO MainThread:4148880 [wandb_init.py:init():813] calling init triggers +2025-09-14 15:29:53,275 INFO MainThread:4148880 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 15:29:53,275 INFO MainThread:4148880 [wandb_init.py:init():854] starting backend +2025-09-14 15:29:53,494 INFO MainThread:4148880 [wandb_init.py:init():857] sending inform_init request +2025-09-14 15:29:53,499 INFO MainThread:4148880 [wandb_init.py:init():865] backend started and connected +2025-09-14 15:29:53,502 INFO MainThread:4148880 [wandb_init.py:init():936] updated telemetry +2025-09-14 15:29:53,513 INFO MainThread:4148880 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 15:29:54,337 INFO MainThread:4148880 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 15:29:54,500 INFO MainThread:4148880 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 15:29:54,500 INFO MainThread:4148880 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 15:29:54,500 INFO MainThread:4148880 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 15:29:54,500 INFO MainThread:4148880 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 15:29:54,503 INFO MainThread:4148880 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 15:30:22,526 INFO MainThread:4148880 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250914_152953-h7gbi509/run-h7gbi509.wandb b/wandb/run-20250914_152953-h7gbi509/run-h7gbi509.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_153740-mntr6bpr/files/output.log b/wandb/run-20250914_153740-mntr6bpr/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7dd707a51391a667e0707c7b29caad5480bb8706 --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/files/output.log @@ -0,0 +1,56 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 245, in read_data + for message in game["history"]: +TypeError: string indices must be integers +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 245, in read_data +[rank0]: for message in game["history"]: +[rank0]: TypeError: string indices must be integers diff --git a/wandb/run-20250914_153740-mntr6bpr/files/requirements.txt b/wandb/run-20250914_153740-mntr6bpr/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_153740-mntr6bpr/files/wandb-metadata.json b/wandb/run-20250914_153740-mntr6bpr/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4e655dc242c12441248983e6223c941cc699aa3b --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/files/wandb-metadata.json @@ -0,0 +1,47 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T07:37:40.225464Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "writerId": "6fssm6j9iqbb67avy1b5yeqw19wlvo70" +} \ No newline at end of file diff --git a/wandb/run-20250914_153740-mntr6bpr/files/wandb-summary.json b/wandb/run-20250914_153740-mntr6bpr/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..7b4fdaac0ef2f0f7b27bbdd0252e8c2048547735 --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":1},"_runtime":1} \ No newline at end of file diff --git a/wandb/run-20250914_153740-mntr6bpr/logs/debug-core.log b/wandb/run-20250914_153740-mntr6bpr/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..f85ac49566ddf24d413403862e5b4994510e0ad8 --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/logs/debug-core.log @@ -0,0 +1,14 @@ +{"time":"2025-09-14T15:37:40.262036642+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp3as27yvq/port-4153177.txt","pid":4153177,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T15:37:40.262325027+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T15:37:40.262984287+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4153177} +{"time":"2025-09-14T15:37:40.262950449+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4153177-4154261-3589761858/socket","Net":"unix"}} +{"time":"2025-09-14T15:37:40.450727428+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T15:37:40.459415135+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"mntr6bpr","id":"1(@)"} +{"time":"2025-09-14T15:37:40.948915731+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"mntr6bpr","id":"1(@)"} +{"time":"2025-09-14T15:37:45.774474936+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T15:37:45.774574071+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T15:37:45.774701578+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T15:37:45.774727602+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T15:37:45.775175493+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4153177-4154261-3589761858/socket","Net":"unix"}} +{"time":"2025-09-14T15:37:46.16316757+08:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write unix /tmp/wandb-4153177-4154261-3589761858/socket->@: use of closed network connection","id":"1(@)"} +{"time":"2025-09-14T15:37:48.594798308+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_153740-mntr6bpr/logs/debug-internal.log b/wandb/run-20250914_153740-mntr6bpr/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a47d65bc4fd2b25bb6a907c225f8f7061a67588f --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T15:37:40.459637278+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T15:37:40.493573196+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T15:37:40.948798768+08:00","level":"INFO","msg":"stream: created new stream","id":"mntr6bpr"} +{"time":"2025-09-14T15:37:40.948901606+08:00","level":"INFO","msg":"stream: started","id":"mntr6bpr"} +{"time":"2025-09-14T15:37:40.948916329+08:00","level":"INFO","msg":"writer: started","stream_id":"mntr6bpr"} +{"time":"2025-09-14T15:37:40.948940209+08:00","level":"INFO","msg":"sender: started","stream_id":"mntr6bpr"} +{"time":"2025-09-14T15:37:40.948975523+08:00","level":"INFO","msg":"handler: started","stream_id":"mntr6bpr"} +{"time":"2025-09-14T15:37:45.774588212+08:00","level":"INFO","msg":"stream: closing","id":"mntr6bpr"} diff --git a/wandb/run-20250914_153740-mntr6bpr/logs/debug.log b/wandb/run-20250914_153740-mntr6bpr/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..b37e2bb317718c5a5b99a0208d1cb0661934f5d2 --- /dev/null +++ b/wandb/run-20250914_153740-mntr6bpr/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 15:37:40,230 INFO MainThread:4153177 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 15:37:40,230 INFO MainThread:4153177 [wandb_setup.py:_flush():81] Configure stats pid to 4153177 +2025-09-14 15:37:40,230 INFO MainThread:4153177 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 15:37:40,230 INFO MainThread:4153177 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 15:37:40,230 INFO MainThread:4153177 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 15:37:40,231 INFO MainThread:4153177 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_153740-mntr6bpr/logs/debug.log +2025-09-14 15:37:40,231 INFO MainThread:4153177 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_153740-mntr6bpr/logs/debug-internal.log +2025-09-14 15:37:40,231 INFO MainThread:4153177 [wandb_init.py:init():813] calling init triggers +2025-09-14 15:37:40,231 INFO MainThread:4153177 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 15:37:40,231 INFO MainThread:4153177 [wandb_init.py:init():854] starting backend +2025-09-14 15:37:40,451 INFO MainThread:4153177 [wandb_init.py:init():857] sending inform_init request +2025-09-14 15:37:40,455 INFO MainThread:4153177 [wandb_init.py:init():865] backend started and connected +2025-09-14 15:37:40,458 INFO MainThread:4153177 [wandb_init.py:init():936] updated telemetry +2025-09-14 15:37:40,469 INFO MainThread:4153177 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 15:37:44,770 INFO MainThread:4153177 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 15:37:44,952 INFO MainThread:4153177 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 15:37:44,952 INFO MainThread:4153177 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 15:37:44,952 INFO MainThread:4153177 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 15:37:44,952 INFO MainThread:4153177 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 15:37:44,954 INFO MainThread:4153177 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 15:37:45,774 INFO wandb-AsyncioManager-main:4153177 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 15:37:45,774 INFO wandb-AsyncioManager-main:4153177 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250914_153740-mntr6bpr/run-mntr6bpr.wandb b/wandb/run-20250914_153740-mntr6bpr/run-mntr6bpr.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_153809-f5bb1kg8/files/config.yaml b/wandb/run-20250914_153809-f5bb1kg8/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3aee62ca745aa2c76c60f872298cfac6cc1097ef --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + 9asqybjnmr7dsjtzh3u0du2yb54b12az: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3584827625472" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T07:38:09.657438Z" + writerId: 9asqybjnmr7dsjtzh3u0du2yb54b12az + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250914_153809-f5bb1kg8/files/output.log b/wandb/run-20250914_153809-f5bb1kg8/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7dd707a51391a667e0707c7b29caad5480bb8706 --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/files/output.log @@ -0,0 +1,56 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 245, in read_data + for message in game["history"]: +TypeError: string indices must be integers +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 221, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 245, in read_data +[rank0]: for message in game["history"]: +[rank0]: TypeError: string indices must be integers diff --git a/wandb/run-20250914_153809-f5bb1kg8/files/requirements.txt b/wandb/run-20250914_153809-f5bb1kg8/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_153809-f5bb1kg8/files/wandb-metadata.json b/wandb/run-20250914_153809-f5bb1kg8/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..9e822d4c4fc8ccdea15387bfec88fa041c89f7b3 --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T07:38:09.657438Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584827625472" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "9asqybjnmr7dsjtzh3u0du2yb54b12az" +} \ No newline at end of file diff --git a/wandb/run-20250914_153809-f5bb1kg8/files/wandb-summary.json b/wandb/run-20250914_153809-f5bb1kg8/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250914_153809-f5bb1kg8/logs/debug-core.log b/wandb/run-20250914_153809-f5bb1kg8/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..4d0d9e7d2eacab94eaf1db223ed01eff0da7d48f --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T15:38:09.69303932+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpt7kjra6q/port-4154526.txt","pid":4154526,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T15:38:09.693389203+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T15:38:09.694316257+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4154526} +{"time":"2025-09-14T15:38:09.694312087+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4154526-4155304-925384313/socket","Net":"unix"}} +{"time":"2025-09-14T15:38:09.883276656+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T15:38:09.894282556+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"f5bb1kg8","id":"1(@)"} +{"time":"2025-09-14T15:38:10.351042155+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"f5bb1kg8","id":"1(@)"} +{"time":"2025-09-14T15:38:12.958064056+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T15:38:12.958151127+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T15:38:12.958133968+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T15:38:12.958346776+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T15:38:12.958473224+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4154526-4155304-925384313/socket","Net":"unix"}} +{"time":"2025-09-14T15:38:14.247948629+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T15:38:14.247977187+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T15:38:14.247983898+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_153809-f5bb1kg8/logs/debug-internal.log b/wandb/run-20250914_153809-f5bb1kg8/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..c8e4775a72d90a73d710164b3b01234297c47586 --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T15:38:09.89467487+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T15:38:09.912847275+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T15:38:10.350848869+08:00","level":"INFO","msg":"stream: created new stream","id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:10.351018138+08:00","level":"INFO","msg":"stream: started","id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:10.351043138+08:00","level":"INFO","msg":"writer: started","stream_id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:10.351059135+08:00","level":"INFO","msg":"sender: started","stream_id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:10.351093095+08:00","level":"INFO","msg":"handler: started","stream_id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:12.958148732+08:00","level":"INFO","msg":"stream: closing","id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:13.999395206+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T15:38:14.246031267+08:00","level":"INFO","msg":"handler: closed","stream_id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:14.246742313+08:00","level":"INFO","msg":"sender: closed","stream_id":"f5bb1kg8"} +{"time":"2025-09-14T15:38:14.246794194+08:00","level":"INFO","msg":"stream: closed","id":"f5bb1kg8"} diff --git a/wandb/run-20250914_153809-f5bb1kg8/logs/debug.log b/wandb/run-20250914_153809-f5bb1kg8/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..9fc06ceef185be7b83184c829455c69078e418bd --- /dev/null +++ b/wandb/run-20250914_153809-f5bb1kg8/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-14 15:38:09,661 INFO MainThread:4154526 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_setup.py:_flush():81] Configure stats pid to 4154526 +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_153809-f5bb1kg8/logs/debug.log +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_153809-f5bb1kg8/logs/debug-internal.log +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_init.py:init():813] calling init triggers +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 15:38:09,662 INFO MainThread:4154526 [wandb_init.py:init():854] starting backend +2025-09-14 15:38:09,883 INFO MainThread:4154526 [wandb_init.py:init():857] sending inform_init request +2025-09-14 15:38:09,888 INFO MainThread:4154526 [wandb_init.py:init():865] backend started and connected +2025-09-14 15:38:09,890 INFO MainThread:4154526 [wandb_init.py:init():936] updated telemetry +2025-09-14 15:38:09,902 INFO MainThread:4154526 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 15:38:12,347 INFO MainThread:4154526 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 15:38:12,532 INFO MainThread:4154526 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 15:38:12,532 INFO MainThread:4154526 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 15:38:12,532 INFO MainThread:4154526 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 15:38:12,532 INFO MainThread:4154526 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 15:38:12,535 INFO MainThread:4154526 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 15:38:12,959 INFO wandb-AsyncioManager-main:4154526 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 15:38:12,959 INFO wandb-AsyncioManager-main:4154526 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_153809-f5bb1kg8/run-f5bb1kg8.wandb b/wandb/run-20250914_153809-f5bb1kg8/run-f5bb1kg8.wandb new file mode 100644 index 0000000000000000000000000000000000000000..d335d6048f7094ba389f854fe2a9097c783b3986 Binary files /dev/null and b/wandb/run-20250914_153809-f5bb1kg8/run-f5bb1kg8.wandb differ diff --git a/wandb/run-20250914_154539-72w1g5l9/files/config.yaml b/wandb/run-20250914_154539-72w1g5l9/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..99606e6299c1ad3f235b4e3f2e85c92c99a3c101 --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 924tsq6rk3yjf4owg1dqbxr5oqnc1ztt: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3584804212736" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T07:45:39.428707Z" + writerId: 924tsq6rk3yjf4owg1dqbxr5oqnc1ztt + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_154539-72w1g5l9/files/output.log b/wandb/run-20250914_154539-72w1g5l9/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..015a7d19e2b37d805d2c83ab00a20263ba6fa08d --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/files/output.log @@ -0,0 +1,12 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 1/527 [00:29<4:18:51, 0.03it/s, v_num=g5l9] +[rank: 0] Received SIGTERM: 15 + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250914_154539-72w1g5l9/files/requirements.txt b/wandb/run-20250914_154539-72w1g5l9/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_154539-72w1g5l9/files/wandb-metadata.json b/wandb/run-20250914_154539-72w1g5l9/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b446b73c4bf02799dbff52d1aca10dac8c56a62b --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T07:45:39.428707Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584804212736" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "924tsq6rk3yjf4owg1dqbxr5oqnc1ztt" +} \ No newline at end of file diff --git a/wandb/run-20250914_154539-72w1g5l9/files/wandb-summary.json b/wandb/run-20250914_154539-72w1g5l9/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..a511638a69bed9bbb6053ba5853a487c3ae2f234 --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":96},"_runtime":96} \ No newline at end of file diff --git a/wandb/run-20250914_154539-72w1g5l9/logs/debug-core.log b/wandb/run-20250914_154539-72w1g5l9/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..6e3539927de1d9d4f46c3a5d93b2aa85b800bb74 --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T15:45:39.464158325+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpp0bw5z7j/port-4158548.txt","pid":4158548,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T15:45:39.464392932+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T15:45:39.465113012+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4158548} +{"time":"2025-09-14T15:45:39.465107093+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4158548-4159175-4028354536/socket","Net":"unix"}} +{"time":"2025-09-14T15:45:39.654480969+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T15:45:39.665067491+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"72w1g5l9","id":"1(@)"} +{"time":"2025-09-14T15:45:40.144347641+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"72w1g5l9","id":"1(@)"} +{"time":"2025-09-14T15:47:16.621678287+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T15:47:16.621867823+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T15:47:16.62196665+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T15:47:16.62199948+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T15:47:16.62268552+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4158548-4159175-4028354536/socket","Net":"unix"}} +{"time":"2025-09-14T15:47:17.869974747+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T15:47:17.870023471+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T15:47:17.870044084+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_154539-72w1g5l9/logs/debug-internal.log b/wandb/run-20250914_154539-72w1g5l9/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..72d2ed63c57f24dc5fc8ec3231835644bd004e08 --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T15:45:39.665366486+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T15:45:39.706489651+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T15:45:40.144254951+08:00","level":"INFO","msg":"stream: created new stream","id":"72w1g5l9"} +{"time":"2025-09-14T15:45:40.144332837+08:00","level":"INFO","msg":"stream: started","id":"72w1g5l9"} +{"time":"2025-09-14T15:45:40.144365167+08:00","level":"INFO","msg":"writer: started","stream_id":"72w1g5l9"} +{"time":"2025-09-14T15:45:40.144375704+08:00","level":"INFO","msg":"sender: started","stream_id":"72w1g5l9"} +{"time":"2025-09-14T15:45:40.14441529+08:00","level":"INFO","msg":"handler: started","stream_id":"72w1g5l9"} +{"time":"2025-09-14T15:47:16.62179501+08:00","level":"INFO","msg":"stream: closing","id":"72w1g5l9"} +{"time":"2025-09-14T15:47:17.213834821+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T15:47:17.867584441+08:00","level":"INFO","msg":"handler: closed","stream_id":"72w1g5l9"} +{"time":"2025-09-14T15:47:17.868523003+08:00","level":"INFO","msg":"sender: closed","stream_id":"72w1g5l9"} +{"time":"2025-09-14T15:47:17.868585294+08:00","level":"INFO","msg":"stream: closed","id":"72w1g5l9"} diff --git a/wandb/run-20250914_154539-72w1g5l9/logs/debug.log b/wandb/run-20250914_154539-72w1g5l9/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..681f02981a3ae04bf7718a71da35b39a5eb432e9 --- /dev/null +++ b/wandb/run-20250914_154539-72w1g5l9/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 15:45:39,432 INFO MainThread:4158548 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_setup.py:_flush():81] Configure stats pid to 4158548 +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_154539-72w1g5l9/logs/debug.log +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_154539-72w1g5l9/logs/debug-internal.log +2025-09-14 15:45:39,433 INFO MainThread:4158548 [wandb_init.py:init():813] calling init triggers +2025-09-14 15:45:39,434 INFO MainThread:4158548 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 15:45:39,434 INFO MainThread:4158548 [wandb_init.py:init():854] starting backend +2025-09-14 15:45:39,653 INFO MainThread:4158548 [wandb_init.py:init():857] sending inform_init request +2025-09-14 15:45:39,658 INFO MainThread:4158548 [wandb_init.py:init():865] backend started and connected +2025-09-14 15:45:39,661 INFO MainThread:4158548 [wandb_init.py:init():936] updated telemetry +2025-09-14 15:45:39,673 INFO MainThread:4158548 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 15:45:40,422 INFO MainThread:4158548 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 15:45:40,591 INFO MainThread:4158548 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 15:45:40,591 INFO MainThread:4158548 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 15:45:40,591 INFO MainThread:4158548 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 15:45:40,591 INFO MainThread:4158548 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 15:45:40,594 INFO MainThread:4158548 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 15:46:08,676 INFO MainThread:4158548 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 15:47:16,621 INFO wandb-AsyncioManager-main:4158548 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 15:47:16,622 INFO wandb-AsyncioManager-main:4158548 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_154539-72w1g5l9/run-72w1g5l9.wandb b/wandb/run-20250914_154539-72w1g5l9/run-72w1g5l9.wandb new file mode 100644 index 0000000000000000000000000000000000000000..77a221dc287d402b4d75a4a84320b76d6020e071 Binary files /dev/null and b/wandb/run-20250914_154539-72w1g5l9/run-72w1g5l9.wandb differ diff --git a/wandb/run-20250914_155155-1xf7liog/files/output.log b/wandb/run-20250914_155155-1xf7liog/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_155155-1xf7liog/files/requirements.txt b/wandb/run-20250914_155155-1xf7liog/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..4457c2cc623c81fc64924de6df2e5e43f446b6e1 --- /dev/null +++ b/wandb/run-20250914_155155-1xf7liog/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_155155-1xf7liog/files/wandb-metadata.json b/wandb/run-20250914_155155-1xf7liog/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6853fdca393682444dc2c8e8d55266cdaa35bcb4 --- /dev/null +++ b/wandb/run-20250914_155155-1xf7liog/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T07:51:55.992880Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584804626432" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "lpdpyaval942vd9z5e8yh9haifaqwviq" +} \ No newline at end of file diff --git a/wandb/run-20250914_155155-1xf7liog/logs/debug-core.log b/wandb/run-20250914_155155-1xf7liog/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..d9e01a990528ae6c4bca907d4e2783fe92f3e131 --- /dev/null +++ b/wandb/run-20250914_155155-1xf7liog/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T15:51:56.061458363+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpzg4hx9ur/port-4161526.txt","pid":4161526,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T15:51:56.061632721+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T15:51:56.062340575+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4161526} +{"time":"2025-09-14T15:51:56.062341792+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4161526-4166912-2871233005/socket","Net":"unix"}} +{"time":"2025-09-14T15:51:56.252147435+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T15:51:56.278409648+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"1xf7liog","id":"1(@)"} +{"time":"2025-09-14T15:51:56.799749145+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"1xf7liog","id":"1(@)"} +{"time":"2025-09-14T15:55:59.978381147+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_155155-1xf7liog/logs/debug-internal.log b/wandb/run-20250914_155155-1xf7liog/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..98761b75fd473ad823d3f13a57970a451cce3c7e --- /dev/null +++ b/wandb/run-20250914_155155-1xf7liog/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T15:51:56.278777957+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T15:51:56.323939706+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T15:51:56.799626139+08:00","level":"INFO","msg":"stream: created new stream","id":"1xf7liog"} +{"time":"2025-09-14T15:51:56.799734925+08:00","level":"INFO","msg":"stream: started","id":"1xf7liog"} +{"time":"2025-09-14T15:51:56.79976595+08:00","level":"INFO","msg":"writer: started","stream_id":"1xf7liog"} +{"time":"2025-09-14T15:51:56.799778055+08:00","level":"INFO","msg":"sender: started","stream_id":"1xf7liog"} +{"time":"2025-09-14T15:51:56.799809644+08:00","level":"INFO","msg":"handler: started","stream_id":"1xf7liog"} diff --git a/wandb/run-20250914_155155-1xf7liog/logs/debug.log b/wandb/run-20250914_155155-1xf7liog/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a6c1b18eee4021e4db92e552a9fc7c097d3e7a56 --- /dev/null +++ b/wandb/run-20250914_155155-1xf7liog/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 15:51:55,999 INFO MainThread:4161526 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 15:51:56,000 INFO MainThread:4161526 [wandb_setup.py:_flush():81] Configure stats pid to 4161526 +2025-09-14 15:51:56,000 INFO MainThread:4161526 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 15:51:56,000 INFO MainThread:4161526 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 15:51:56,001 INFO MainThread:4161526 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 15:51:56,001 INFO MainThread:4161526 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_155155-1xf7liog/logs/debug.log +2025-09-14 15:51:56,002 INFO MainThread:4161526 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_155155-1xf7liog/logs/debug-internal.log +2025-09-14 15:51:56,002 INFO MainThread:4161526 [wandb_init.py:init():813] calling init triggers +2025-09-14 15:51:56,002 INFO MainThread:4161526 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 15:51:56,003 INFO MainThread:4161526 [wandb_init.py:init():854] starting backend +2025-09-14 15:51:56,253 INFO MainThread:4161526 [wandb_init.py:init():857] sending inform_init request +2025-09-14 15:51:56,263 INFO MainThread:4161526 [wandb_init.py:init():865] backend started and connected +2025-09-14 15:51:56,275 INFO MainThread:4161526 [wandb_init.py:init():936] updated telemetry +2025-09-14 15:51:56,300 INFO MainThread:4161526 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 15:51:57,130 INFO MainThread:4161526 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 15:51:57,498 INFO MainThread:4161526 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 15:51:57,500 INFO MainThread:4161526 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 15:51:57,500 INFO MainThread:4161526 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 15:51:57,501 INFO MainThread:4161526 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 15:51:57,508 INFO MainThread:4161526 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_155155-1xf7liog/run-1xf7liog.wandb b/wandb/run-20250914_155155-1xf7liog/run-1xf7liog.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_161326-16xxo3uy/files/output.log b/wandb/run-20250914_161326-16xxo3uy/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_161326-16xxo3uy/files/requirements.txt b/wandb/run-20250914_161326-16xxo3uy/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..4457c2cc623c81fc64924de6df2e5e43f446b6e1 --- /dev/null +++ b/wandb/run-20250914_161326-16xxo3uy/files/requirements.txt @@ -0,0 +1,141 @@ +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_161326-16xxo3uy/files/wandb-metadata.json b/wandb/run-20250914_161326-16xxo3uy/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..1f7301ee70f5ea70c415eeff8a81f02f70997891 --- /dev/null +++ b/wandb/run-20250914_161326-16xxo3uy/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T08:13:26.282981Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.n_traj_eval=4", + "--data.base_model=Qwen3-14B", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.limit_val_batches=0", + "--trainer.val_check_interval=null", + "--trainer.enable_model_summary=false" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584806035456" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ap1egcwxg9j51dzy2vf1i1gd762dq4h6" +} \ No newline at end of file diff --git a/wandb/run-20250914_161326-16xxo3uy/logs/debug-core.log b/wandb/run-20250914_161326-16xxo3uy/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..0410213281cfad0c9ccc612ca49ec36eed4581fd --- /dev/null +++ b/wandb/run-20250914_161326-16xxo3uy/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-14T16:13:26.349609705+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpvpyfkvdz/port-4175701.txt","pid":4175701,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T16:13:26.349873252+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat $HOME/tmp: no such file or directory"} +{"time":"2025-09-14T16:13:26.350699464+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4175701} +{"time":"2025-09-14T16:13:26.350696741+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4175701-4188514-3007398483/socket","Net":"unix"}} +{"time":"2025-09-14T16:13:26.540557526+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T16:13:26.568336542+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"16xxo3uy","id":"1(@)"} +{"time":"2025-09-14T16:13:27.018985437+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"16xxo3uy","id":"1(@)"} +{"time":"2025-09-14T16:14:02.180843483+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250914_161326-16xxo3uy/logs/debug-internal.log b/wandb/run-20250914_161326-16xxo3uy/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a3ee3bb8e06036ee6038e2919230eb760d0e6fc7 --- /dev/null +++ b/wandb/run-20250914_161326-16xxo3uy/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-14T16:13:26.568510772+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T16:13:26.587375708+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T16:13:27.018850257+08:00","level":"INFO","msg":"stream: created new stream","id":"16xxo3uy"} +{"time":"2025-09-14T16:13:27.018970017+08:00","level":"INFO","msg":"stream: started","id":"16xxo3uy"} +{"time":"2025-09-14T16:13:27.019010063+08:00","level":"INFO","msg":"writer: started","stream_id":"16xxo3uy"} +{"time":"2025-09-14T16:13:27.019203392+08:00","level":"INFO","msg":"sender: started","stream_id":"16xxo3uy"} +{"time":"2025-09-14T16:13:27.019239477+08:00","level":"INFO","msg":"handler: started","stream_id":"16xxo3uy"} diff --git a/wandb/run-20250914_161326-16xxo3uy/logs/debug.log b/wandb/run-20250914_161326-16xxo3uy/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..094d954624ed32ffefa4cf86e747c9163c759ca4 --- /dev/null +++ b/wandb/run-20250914_161326-16xxo3uy/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-14 16:13:26,291 INFO MainThread:4175701 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 16:13:26,291 INFO MainThread:4175701 [wandb_setup.py:_flush():81] Configure stats pid to 4175701 +2025-09-14 16:13:26,292 INFO MainThread:4175701 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 16:13:26,292 INFO MainThread:4175701 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 16:13:26,292 INFO MainThread:4175701 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 16:13:26,293 INFO MainThread:4175701 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_161326-16xxo3uy/logs/debug.log +2025-09-14 16:13:26,293 INFO MainThread:4175701 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_161326-16xxo3uy/logs/debug-internal.log +2025-09-14 16:13:26,294 INFO MainThread:4175701 [wandb_init.py:init():813] calling init triggers +2025-09-14 16:13:26,294 INFO MainThread:4175701 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 16:13:26,294 INFO MainThread:4175701 [wandb_init.py:init():854] starting backend +2025-09-14 16:13:26,541 INFO MainThread:4175701 [wandb_init.py:init():857] sending inform_init request +2025-09-14 16:13:26,552 INFO MainThread:4175701 [wandb_init.py:init():865] backend started and connected +2025-09-14 16:13:26,565 INFO MainThread:4175701 [wandb_init.py:init():936] updated telemetry +2025-09-14 16:13:26,591 INFO MainThread:4175701 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 16:13:27,382 INFO MainThread:4175701 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 16:13:27,825 INFO MainThread:4175701 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 16:13:27,825 INFO MainThread:4175701 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 16:13:27,826 INFO MainThread:4175701 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 16:13:27,826 INFO MainThread:4175701 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 16:13:27,833 INFO MainThread:4175701 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250914_161326-16xxo3uy/run-16xxo3uy.wandb b/wandb/run-20250914_161326-16xxo3uy/run-16xxo3uy.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250914_162826-00o3oa6m/files/output.log b/wandb/run-20250914_162826-00o3oa6m/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..06bf5190a0959761ef088125c297d213bf0a643c --- /dev/null +++ b/wandb/run-20250914_162826-00o3oa6m/files/output.log @@ -0,0 +1,10 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (41) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/41 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 493, in training_step + print(len(batch[0])) +KeyError: 0 +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 493, in training_step +[rank0]: print(len(batch[0])) +[rank0]: KeyError: 0 diff --git a/wandb/run-20250914_163643-s7a9m43m/files/requirements.txt b/wandb/run-20250914_163643-s7a9m43m/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250914_163643-s7a9m43m/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_163643-s7a9m43m/files/wandb-metadata.json b/wandb/run-20250914_163643-s7a9m43m/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e7268f2d881dda29fce6c48890e999cc18db2b9a --- /dev/null +++ b/wandb/run-20250914_163643-s7a9m43m/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T08:36:43.709292Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584805249024" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "l4c8qa6fk24g4crz1wulkd65xqyj1c8n" +} \ No newline at end of file diff --git a/wandb/run-20250914_163643-s7a9m43m/files/wandb-summary.json b/wandb/run-20250914_163643-s7a9m43m/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..7c964515f852314220af92c8082d4fb2309b0bbd --- /dev/null +++ b/wandb/run-20250914_163643-s7a9m43m/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":34,"_wandb":{"runtime":34}} \ No newline at end of file diff --git a/wandb/run-20250914_163643-s7a9m43m/logs/debug-core.log b/wandb/run-20250914_163643-s7a9m43m/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..aef1b240f36396f7fb779dc1afb2e7e1097265ea --- /dev/null +++ b/wandb/run-20250914_163643-s7a9m43m/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T16:36:43.736044323+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp1zpcb97e/port-19035.txt","pid":19035,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T16:36:43.736268081+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T16:36:43.737270849+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-19035-19727-1132563347/socket","Net":"unix"}} +{"time":"2025-09-14T16:36:43.737303581+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":19035} +{"time":"2025-09-14T16:36:43.922800227+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T16:36:43.930304181+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"s7a9m43m","id":"1(@)"} +{"time":"2025-09-14T16:36:44.40829579+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"s7a9m43m","id":"1(@)"} +{"time":"2025-09-14T16:37:19.032407821+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T16:37:19.032473977+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T16:37:19.032464587+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T16:37:19.032557816+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T16:37:19.032668613+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-19035-19727-1132563347/socket","Net":"unix"}} +{"time":"2025-09-14T16:37:19.906450675+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T16:37:19.90649615+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T16:37:19.906515516+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_163643-s7a9m43m/logs/debug-internal.log b/wandb/run-20250914_163643-s7a9m43m/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..805102fcff7fdf67244fd313ab4e41270f442bf6 --- /dev/null +++ b/wandb/run-20250914_163643-s7a9m43m/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T16:36:43.930455143+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T16:36:43.952579242+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T16:36:44.408223545+08:00","level":"INFO","msg":"stream: created new stream","id":"s7a9m43m"} +{"time":"2025-09-14T16:36:44.408289718+08:00","level":"INFO","msg":"stream: started","id":"s7a9m43m"} +{"time":"2025-09-14T16:36:44.408340807+08:00","level":"INFO","msg":"writer: started","stream_id":"s7a9m43m"} +{"time":"2025-09-14T16:36:44.408428159+08:00","level":"INFO","msg":"sender: started","stream_id":"s7a9m43m"} +{"time":"2025-09-14T16:36:44.408347859+08:00","level":"INFO","msg":"handler: started","stream_id":"s7a9m43m"} +{"time":"2025-09-14T16:37:19.032481562+08:00","level":"INFO","msg":"stream: closing","id":"s7a9m43m"} +{"time":"2025-09-14T16:37:19.642735376+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T16:37:19.904538045+08:00","level":"INFO","msg":"handler: closed","stream_id":"s7a9m43m"} +{"time":"2025-09-14T16:37:19.905162057+08:00","level":"INFO","msg":"sender: closed","stream_id":"s7a9m43m"} +{"time":"2025-09-14T16:37:19.905219091+08:00","level":"INFO","msg":"stream: closed","id":"s7a9m43m"} diff --git a/wandb/run-20250914_163643-s7a9m43m/logs/debug.log b/wandb/run-20250914_163643-s7a9m43m/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6f0239c18d41427e79ce03ac6161e361012b8c5d --- /dev/null +++ b/wandb/run-20250914_163643-s7a9m43m/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_setup.py:_flush():81] Configure stats pid to 19035 +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_163643-s7a9m43m/logs/debug.log +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_163643-s7a9m43m/logs/debug-internal.log +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_init.py:init():813] calling init triggers +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 16:36:43,711 INFO MainThread:19035 [wandb_init.py:init():854] starting backend +2025-09-14 16:36:43,923 INFO MainThread:19035 [wandb_init.py:init():857] sending inform_init request +2025-09-14 16:36:43,927 INFO MainThread:19035 [wandb_init.py:init():865] backend started and connected +2025-09-14 16:36:43,930 INFO MainThread:19035 [wandb_init.py:init():936] updated telemetry +2025-09-14 16:36:43,941 INFO MainThread:19035 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 16:36:44,754 INFO MainThread:19035 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 16:36:44,940 INFO MainThread:19035 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 16:36:44,940 INFO MainThread:19035 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 16:36:44,940 INFO MainThread:19035 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 16:36:44,940 INFO MainThread:19035 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 16:36:44,943 INFO MainThread:19035 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 16:37:12,701 INFO MainThread:19035 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 16:37:19,032 INFO wandb-AsyncioManager-main:19035 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 16:37:19,032 INFO wandb-AsyncioManager-main:19035 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_163643-s7a9m43m/run-s7a9m43m.wandb b/wandb/run-20250914_163643-s7a9m43m/run-s7a9m43m.wandb new file mode 100644 index 0000000000000000000000000000000000000000..caccd8bd88382ca427a333b1388c9f25acd8b58d Binary files /dev/null and b/wandb/run-20250914_163643-s7a9m43m/run-s7a9m43m.wandb differ diff --git a/wandb/run-20250914_163815-xiygbdu2/files/config.yaml b/wandb/run-20250914_163815-xiygbdu2/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..43063ef4001ac47654f6d057d0869146f78b5440 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + eiulaw5nxs5x453okkjau8hj9pkcd5ss: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3584805593088" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T08:38:15.865636Z" + writerId: eiulaw5nxs5x453okkjau8hj9pkcd5ss + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_163815-xiygbdu2/files/output.log b/wandb/run-20250914_163815-xiygbdu2/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7cac41843ca88dc6558f08b3af22682d3251a6d0 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/files/output.log @@ -0,0 +1,125 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/527 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 493, in training_step + print(batch[0]) +KeyError: 0 +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 24, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 13, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 493, in training_step +[rank0]: print(batch[0]) +[rank0]: KeyError: 0 diff --git a/wandb/run-20250914_163815-xiygbdu2/files/requirements.txt b/wandb/run-20250914_163815-xiygbdu2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250914_163815-xiygbdu2/files/wandb-metadata.json b/wandb/run-20250914_163815-xiygbdu2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..d4cb4b7f6f5cd2885171f098faa085a2b14171d3 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-14T08:38:15.865636Z", + "args": [ + "fit", + "--data=RSAGame", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=16", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=RSAGame-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3584805593088" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "eiulaw5nxs5x453okkjau8hj9pkcd5ss" +} \ No newline at end of file diff --git a/wandb/run-20250914_163815-xiygbdu2/files/wandb-summary.json b/wandb/run-20250914_163815-xiygbdu2/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..0a14c801ab848d9af1a7a6f0f9b20080bf02bdb1 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":34},"_runtime":34} \ No newline at end of file diff --git a/wandb/run-20250914_163815-xiygbdu2/logs/debug-core.log b/wandb/run-20250914_163815-xiygbdu2/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..6391212f08682b40610d4ce2eb7f77daed31aa08 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-14T16:38:15.900448243+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpxg1o9d7c/port-21398.txt","pid":21398,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-14T16:38:15.900665177+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-14T16:38:15.901436776+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":21398} +{"time":"2025-09-14T16:38:15.901439335+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-21398-21941-1849781022/socket","Net":"unix"}} +{"time":"2025-09-14T16:38:16.088494069+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-14T16:38:16.095781737+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"xiygbdu2","id":"1(@)"} +{"time":"2025-09-14T16:38:16.560812653+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"xiygbdu2","id":"1(@)"} +{"time":"2025-09-14T16:38:51.518835765+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-14T16:38:51.519203568+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-14T16:38:51.519636173+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-14T16:38:51.519920059+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-14T16:38:51.520205477+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-21398-21941-1849781022/socket","Net":"unix"}} +{"time":"2025-09-14T16:38:52.551904396+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T16:38:52.551943736+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T16:38:52.551962443+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_163815-xiygbdu2/logs/debug-internal.log b/wandb/run-20250914_163815-xiygbdu2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..be4b1ecbcbee5f6e0ae9153eb1b3044d344358b8 --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T16:38:16.095952966+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T16:38:16.118909832+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T16:38:16.560388773+08:00","level":"INFO","msg":"stream: created new stream","id":"xiygbdu2"} +{"time":"2025-09-14T16:38:16.560796829+08:00","level":"INFO","msg":"stream: started","id":"xiygbdu2"} +{"time":"2025-09-14T16:38:16.560846613+08:00","level":"INFO","msg":"writer: started","stream_id":"xiygbdu2"} +{"time":"2025-09-14T16:38:16.560960734+08:00","level":"INFO","msg":"sender: started","stream_id":"xiygbdu2"} +{"time":"2025-09-14T16:38:16.56100719+08:00","level":"INFO","msg":"handler: started","stream_id":"xiygbdu2"} +{"time":"2025-09-14T16:38:51.518994253+08:00","level":"INFO","msg":"stream: closing","id":"xiygbdu2"} +{"time":"2025-09-14T16:38:52.268606676+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T16:38:52.549551227+08:00","level":"INFO","msg":"handler: closed","stream_id":"xiygbdu2"} +{"time":"2025-09-14T16:38:52.550302932+08:00","level":"INFO","msg":"sender: closed","stream_id":"xiygbdu2"} +{"time":"2025-09-14T16:38:52.550376128+08:00","level":"INFO","msg":"stream: closed","id":"xiygbdu2"} diff --git a/wandb/run-20250914_163815-xiygbdu2/logs/debug.log b/wandb/run-20250914_163815-xiygbdu2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..810d17aff6a37a78560e6dcc955687448afac26e --- /dev/null +++ b/wandb/run-20250914_163815-xiygbdu2/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 16:38:15,869 INFO MainThread:21398 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_setup.py:_flush():81] Configure stats pid to 21398 +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_163815-xiygbdu2/logs/debug.log +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_163815-xiygbdu2/logs/debug-internal.log +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_init.py:init():813] calling init triggers +2025-09-14 16:38:15,870 INFO MainThread:21398 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 16:38:15,871 INFO MainThread:21398 [wandb_init.py:init():854] starting backend +2025-09-14 16:38:16,088 INFO MainThread:21398 [wandb_init.py:init():857] sending inform_init request +2025-09-14 16:38:16,093 INFO MainThread:21398 [wandb_init.py:init():865] backend started and connected +2025-09-14 16:38:16,096 INFO MainThread:21398 [wandb_init.py:init():936] updated telemetry +2025-09-14 16:38:16,106 INFO MainThread:21398 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 16:38:16,955 INFO MainThread:21398 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 16:38:17,123 INFO MainThread:21398 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 16:38:17,123 INFO MainThread:21398 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 16:38:17,123 INFO MainThread:21398 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 16:38:17,123 INFO MainThread:21398 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 16:38:17,125 INFO MainThread:21398 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 16:38:45,167 INFO MainThread:21398 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 16:38:51,522 INFO wandb-AsyncioManager-main:21398 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 16:38:51,523 INFO wandb-AsyncioManager-main:21398 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250914_163815-xiygbdu2/run-xiygbdu2.wandb b/wandb/run-20250914_163815-xiygbdu2/run-xiygbdu2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..2e756c9ff64d68046af1342bce3bae03ad198f1a Binary files /dev/null and b/wandb/run-20250914_163815-xiygbdu2/run-xiygbdu2.wandb differ diff --git a/wandb/run-20250914_163945-hvkinr0d/files/config.yaml b/wandb/run-20250914_163945-hvkinr0d/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..abb699008314c3ee4416c1fd87455e0c3879473c --- /dev/null +++ b/wandb/run-20250914_163945-hvkinr0d/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + cacj31ps9mvctpaz63814ft5rf71puaa: + args: + - fit + - --data=RSAGame + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=RSAGame-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_rsa + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3584805740544" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T08:39:45.147297Z" + writerId: cacj31ps9mvctpaz63814ft5rf71puaa + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_rsa/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_163945-hvkinr0d/files/output.log b/wandb/run-20250914_163945-hvkinr0d/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b46a802d3685ea5c8beb7f7c9aa3afcf612f5f24 --- /dev/null +++ b/wandb/run-20250914_163945-hvkinr0d/files/output.log @@ -0,0 +1,10 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/527 [00:00@: use of closed network connection","id":"1(@)"} +{"time":"2025-09-14T16:43:07.430341265+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-14T16:43:07.430398082+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-14T16:43:07.430438087+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250914_164134-ho7inuxb/logs/debug-internal.log b/wandb/run-20250914_164134-ho7inuxb/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..478cd5a7a90838761f4adabdb59b895e21321e70 --- /dev/null +++ b/wandb/run-20250914_164134-ho7inuxb/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-14T16:41:34.601369148+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-14T16:41:34.633193395+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-14T16:41:35.105697133+08:00","level":"INFO","msg":"stream: created new stream","id":"ho7inuxb"} +{"time":"2025-09-14T16:41:35.105791288+08:00","level":"INFO","msg":"stream: started","id":"ho7inuxb"} +{"time":"2025-09-14T16:41:35.105822854+08:00","level":"INFO","msg":"writer: started","stream_id":"ho7inuxb"} +{"time":"2025-09-14T16:41:35.105839396+08:00","level":"INFO","msg":"sender: started","stream_id":"ho7inuxb"} +{"time":"2025-09-14T16:41:35.105874117+08:00","level":"INFO","msg":"handler: started","stream_id":"ho7inuxb"} +{"time":"2025-09-14T16:43:05.758443852+08:00","level":"INFO","msg":"stream: closing","id":"ho7inuxb"} +{"time":"2025-09-14T16:43:07.116901732+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-14T16:43:07.428759177+08:00","level":"INFO","msg":"handler: closed","stream_id":"ho7inuxb"} +{"time":"2025-09-14T16:43:07.429647336+08:00","level":"INFO","msg":"sender: closed","stream_id":"ho7inuxb"} +{"time":"2025-09-14T16:43:07.429739643+08:00","level":"INFO","msg":"stream: closed","id":"ho7inuxb"} diff --git a/wandb/run-20250914_164134-ho7inuxb/logs/debug.log b/wandb/run-20250914_164134-ho7inuxb/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a6677b75566fffb8e60191c05d63e391e70fd44a --- /dev/null +++ b/wandb/run-20250914_164134-ho7inuxb/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-14 16:41:34,372 INFO MainThread:26482 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-14 16:41:34,372 INFO MainThread:26482 [wandb_setup.py:_flush():81] Configure stats pid to 26482 +2025-09-14 16:41:34,372 INFO MainThread:26482 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-14 16:41:34,372 INFO MainThread:26482 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-14 16:41:34,372 INFO MainThread:26482 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-14 16:41:34,373 INFO MainThread:26482 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250914_164134-ho7inuxb/logs/debug.log +2025-09-14 16:41:34,373 INFO MainThread:26482 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250914_164134-ho7inuxb/logs/debug-internal.log +2025-09-14 16:41:34,373 INFO MainThread:26482 [wandb_init.py:init():813] calling init triggers +2025-09-14 16:41:34,373 INFO MainThread:26482 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-14 16:41:34,373 INFO MainThread:26482 [wandb_init.py:init():854] starting backend +2025-09-14 16:41:34,593 INFO MainThread:26482 [wandb_init.py:init():857] sending inform_init request +2025-09-14 16:41:34,595 INFO MainThread:26482 [wandb_init.py:init():865] backend started and connected +2025-09-14 16:41:34,596 INFO MainThread:26482 [wandb_init.py:init():936] updated telemetry +2025-09-14 16:41:34,603 INFO MainThread:26482 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-14 16:41:35,411 INFO MainThread:26482 [wandb_init.py:init():1011] starting run threads in backend +2025-09-14 16:41:35,604 INFO MainThread:26482 [wandb_run.py:_console_start():2506] atexit reg +2025-09-14 16:41:35,604 INFO MainThread:26482 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-14 16:41:35,605 INFO MainThread:26482 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-14 16:41:35,605 INFO MainThread:26482 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-14 16:41:35,607 INFO MainThread:26482 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-14 16:42:59,547 INFO MainThread:26482 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 16, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-14 16:43:05,758 INFO wandb-AsyncioManager-main:26482 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-14 16:43:05,758 INFO wandb-AsyncioManager-main:26482 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250914_164134-ho7inuxb/run-ho7inuxb.wandb b/wandb/run-20250914_164134-ho7inuxb/run-ho7inuxb.wandb new file mode 100644 index 0000000000000000000000000000000000000000..dcc0989e579d6f33cb08d44e4c01f76fb49b8927 Binary files /dev/null and b/wandb/run-20250914_164134-ho7inuxb/run-ho7inuxb.wandb differ diff --git a/wandb/run-20250914_164331-juirkm35/files/config.yaml b/wandb/run-20250914_164331-juirkm35/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..aff4f2350c55e343b2ab6fda8ce777fa930921d0 --- /dev/null +++ b/wandb/run-20250914_164331-juirkm35/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + p2eoqu6oawmlwc7fhp9apy541c958ypw: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=16 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3584806543360" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-14T08:43:31.797283Z" + writerId: p2eoqu6oawmlwc7fhp9apy541c958ypw + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 16 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250914_164331-juirkm35/files/output.log b/wandb/run-20250914_164331-juirkm35/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8ef3ac157256332a828c69087755be784e37c990 --- /dev/null +++ b/wandb/run-20250914_164331-juirkm35/files/output.log @@ -0,0 +1,11 @@ + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 473, numel = 3938312 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/connectors/data_connector.py:433: The 'train_dataloader' does not have many workers which may be a bottleneck. Consider increasing the value of the `num_workers` argument` to `num_workers=15` in the `DataLoader` to improve performance. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:310: The number of training batches (32) is smaller than the logging interval Trainer(log_every_n_steps=50). Set a lower value for log_every_n_steps if you want to see logs for the training epoch. +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 879 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/32 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 21, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 500, in training_step + self.actor_current_backward_step += 1 + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +torch.AcceleratorError: CUDA error: out of memory +Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. + +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 32, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 21, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 500, in training_step +[rank0]: self.actor_current_backward_step += 1 +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: torch.AcceleratorError: CUDA error: out of memory +[rank0]: Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. diff --git a/wandb/run-20250915_111603-tnh8ytpw/files/requirements.txt b/wandb/run-20250915_111603-tnh8ytpw/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_111603-tnh8ytpw/files/wandb-metadata.json b/wandb/run-20250915_111603-tnh8ytpw/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c5896ecb391cf03caf4b2e204885b44be505edd1 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T03:16:03.066141Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577975308288" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "qcira9hfp6ufx7mg5091jytk3xc2b8un" +} \ No newline at end of file diff --git a/wandb/run-20250915_111603-tnh8ytpw/files/wandb-summary.json b/wandb/run-20250915_111603-tnh8ytpw/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..46fa5848936eb4f4c187dfd6bd93d3a4dc1460c9 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/files/wandb-summary.json @@ -0,0 +1 @@ +{"critic/q1.loss":0.0530548095703125,"actor/factor.max":2.7470703125,"actor/advantages.min":0.631317138671875,"actor/log_prob.mean":-109.59375,"_step":12,"critic/v2.max":-0.07733154296875,"critic/v2.mean":-0.08087158203125,"critic/q1.mean":0.728759765625,"actor/loss":31.27734375,"actor/factor.min":2.0341796875,"_timestamp":1.7579164794544425e+09,"actor/log_prob.min":-170.875,"critic/v1.loss":1.2830855666834395e-05,"_wandb":{"runtime":10343},"critic/v2.loss":1.1017918041034136e-05,"critic/v2.min":-0.0848388671875,"_runtime":10343,"actor/factor.mean":2.390625,"critic/v1.min":0.0155792236328125,"actor/advantages.max":0.98779296875,"critic/target_q1.max":0.014190673828125,"critic/q1.min":0.547698974609375,"critic/target_q2.max":-0.0875244140625,"critic/target_q1.mean":0.01311492919921875,"critic/q2.min":0.6136856079101562,"critic/q2.mean":0.783447265625,"actor/log_prob.max":-48.1953125,"critic/target_q2.min":-0.08990478515625,"critic/q2.loss":0.053287506103515625,"epoch":0,"critic/target_q1.min":0.01207733154296875,"critic/v1.max":0.02313232421875,"actor/advantages.mean":0.80859375,"critic/v1.mean":0.01934051513671875,"trainer/global_step":649,"critic/q2.max":0.952880859375,"critic/q1.max":0.908935546875,"critic/target_q2.mean":-0.08880615234375} \ No newline at end of file diff --git a/wandb/run-20250915_111603-tnh8ytpw/logs/debug-core.log b/wandb/run-20250915_111603-tnh8ytpw/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..f230d490dfff3b9ca8e64032d00a547279741789 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T11:16:03.16922403+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpb61qy1w3/port-803053.txt","pid":803053,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T11:16:03.169472841+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T11:16:03.171717578+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":803053} +{"time":"2025-09-15T11:16:03.171687994+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-803053-803509-4065321074/socket","Net":"unix"}} +{"time":"2025-09-15T11:16:03.357844156+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T11:16:03.384741098+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"tnh8ytpw","id":"1(@)"} +{"time":"2025-09-15T11:16:04.7006821+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"tnh8ytpw","id":"1(@)"} +{"time":"2025-09-15T14:08:28.480807024+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T14:08:28.480947136+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T14:08:28.480949132+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T14:08:28.481121804+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T14:08:28.48121177+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-803053-803509-4065321074/socket","Net":"unix"}} +{"time":"2025-09-15T14:08:29.50693409+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T14:08:29.506994842+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T14:08:29.507032968+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_111603-tnh8ytpw/logs/debug-internal.log b/wandb/run-20250915_111603-tnh8ytpw/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b62d94387958a3c54dfa46e3b083c5e218d9c7a0 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T11:16:03.385000919+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T11:16:03.501141114+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T11:16:04.700546591+08:00","level":"INFO","msg":"stream: created new stream","id":"tnh8ytpw"} +{"time":"2025-09-15T11:16:04.700665535+08:00","level":"INFO","msg":"stream: started","id":"tnh8ytpw"} +{"time":"2025-09-15T11:16:04.70080602+08:00","level":"INFO","msg":"writer: started","stream_id":"tnh8ytpw"} +{"time":"2025-09-15T11:16:04.700925658+08:00","level":"INFO","msg":"sender: started","stream_id":"tnh8ytpw"} +{"time":"2025-09-15T11:16:04.701037232+08:00","level":"INFO","msg":"handler: started","stream_id":"tnh8ytpw"} +{"time":"2025-09-15T14:08:28.480923078+08:00","level":"INFO","msg":"stream: closing","id":"tnh8ytpw"} +{"time":"2025-09-15T14:08:29.227589+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T14:08:29.504971614+08:00","level":"INFO","msg":"handler: closed","stream_id":"tnh8ytpw"} +{"time":"2025-09-15T14:08:29.506058389+08:00","level":"INFO","msg":"sender: closed","stream_id":"tnh8ytpw"} +{"time":"2025-09-15T14:08:29.5061442+08:00","level":"INFO","msg":"stream: closed","id":"tnh8ytpw"} diff --git a/wandb/run-20250915_111603-tnh8ytpw/logs/debug.log b/wandb/run-20250915_111603-tnh8ytpw/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..ed084e0ef18b14e89ef3393b8839f8e3d4dfeb35 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-15 11:16:03,070 INFO MainThread:803053 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 11:16:03,070 INFO MainThread:803053 [wandb_setup.py:_flush():81] Configure stats pid to 803053 +2025-09-15 11:16:03,070 INFO MainThread:803053 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 11:16:03,070 INFO MainThread:803053 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 11:16:03,070 INFO MainThread:803053 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 11:16:03,070 INFO MainThread:803053 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_111603-tnh8ytpw/logs/debug.log +2025-09-15 11:16:03,071 INFO MainThread:803053 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_111603-tnh8ytpw/logs/debug-internal.log +2025-09-15 11:16:03,071 INFO MainThread:803053 [wandb_init.py:init():813] calling init triggers +2025-09-15 11:16:03,071 INFO MainThread:803053 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 11:16:03,071 INFO MainThread:803053 [wandb_init.py:init():854] starting backend +2025-09-15 11:16:03,358 INFO MainThread:803053 [wandb_init.py:init():857] sending inform_init request +2025-09-15 11:16:03,378 INFO MainThread:803053 [wandb_init.py:init():865] backend started and connected +2025-09-15 11:16:03,381 INFO MainThread:803053 [wandb_init.py:init():936] updated telemetry +2025-09-15 11:16:03,393 INFO MainThread:803053 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 11:16:05,062 INFO MainThread:803053 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 11:16:05,452 INFO MainThread:803053 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 11:16:05,453 INFO MainThread:803053 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 11:16:05,453 INFO MainThread:803053 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 11:16:05,453 INFO MainThread:803053 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 11:16:05,485 INFO MainThread:803053 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 11:17:44,460 INFO MainThread:803053 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 8, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-15 14:08:28,482 INFO wandb-AsyncioManager-main:803053 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 14:08:28,482 INFO wandb-AsyncioManager-main:803053 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_111603-tnh8ytpw/run-tnh8ytpw.wandb b/wandb/run-20250915_111603-tnh8ytpw/run-tnh8ytpw.wandb new file mode 100644 index 0000000000000000000000000000000000000000..f8ea9c6ecd18743f228e6a0d223136408aba4218 --- /dev/null +++ b/wandb/run-20250915_111603-tnh8ytpw/run-tnh8ytpw.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59c6ee64612bd23cb71626f08a8c8aaf87a3c737ca2a47607b4fdc8e6bffec59 +size 555640 diff --git a/wandb/run-20250915_170247-9rluwifr/files/output.log b/wandb/run-20250915_170247-9rluwifr/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250915_170247-9rluwifr/files/requirements.txt b/wandb/run-20250915_170247-9rluwifr/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_170247-9rluwifr/files/wandb-metadata.json b/wandb/run-20250915_170247-9rluwifr/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ec3dafaff16bc9ba615ac777225959c9c2d48180 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T09:02:47.297813Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3578274320384" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "8rjg7jwt5nvmk7ua6rxsh6d9ziolof37" +} \ No newline at end of file diff --git a/wandb/run-20250915_170247-9rluwifr/files/wandb-summary.json b/wandb/run-20250915_170247-9rluwifr/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250915_170247-9rluwifr/logs/debug-core.log b/wandb/run-20250915_170247-9rluwifr/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..dc8631a8b0ec13b83194ff9824941688378482f7 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-15T17:02:47.39078949+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpj937ugzy/port-1495353.txt","pid":1495353,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T17:02:47.391080029+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T17:02:47.39209796+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":1495353} +{"time":"2025-09-15T17:02:47.39208202+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-1495353-1497868-1763925040/socket","Net":"unix"}} +{"time":"2025-09-15T17:02:47.579896754+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T17:02:47.605679614+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"9rluwifr","id":"1(@)"} +{"time":"2025-09-15T17:02:48.117551692+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"9rluwifr","id":"1(@)"} +{"time":"2025-09-15T17:02:50.624139286+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T17:02:50.624319348+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T17:02:50.624328906+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T17:02:50.624473868+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T17:02:50.624636429+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-1495353-1497868-1763925040/socket","Net":"unix"}} +{"time":"2025-09-15T17:02:51.207923773+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250915_170247-9rluwifr/logs/debug-internal.log b/wandb/run-20250915_170247-9rluwifr/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..6d398f270d1967f89a929f800b4f7cba0e0291d2 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-15T17:02:47.605872794+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T17:02:47.62245803+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T17:02:48.11742384+08:00","level":"INFO","msg":"stream: created new stream","id":"9rluwifr"} +{"time":"2025-09-15T17:02:48.117536204+08:00","level":"INFO","msg":"stream: started","id":"9rluwifr"} +{"time":"2025-09-15T17:02:48.117567229+08:00","level":"INFO","msg":"sender: started","stream_id":"9rluwifr"} +{"time":"2025-09-15T17:02:48.117562512+08:00","level":"INFO","msg":"writer: started","stream_id":"9rluwifr"} +{"time":"2025-09-15T17:02:48.117604269+08:00","level":"INFO","msg":"handler: started","stream_id":"9rluwifr"} +{"time":"2025-09-15T17:02:50.624313968+08:00","level":"INFO","msg":"stream: closing","id":"9rluwifr"} diff --git a/wandb/run-20250915_170247-9rluwifr/logs/debug.log b/wandb/run-20250915_170247-9rluwifr/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..89df2ed06998136f08ec661c0bf92f2d496913c5 --- /dev/null +++ b/wandb/run-20250915_170247-9rluwifr/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 17:02:47,301 INFO MainThread:1495353 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 17:02:47,301 INFO MainThread:1495353 [wandb_setup.py:_flush():81] Configure stats pid to 1495353 +2025-09-15 17:02:47,301 INFO MainThread:1495353 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 17:02:47,301 INFO MainThread:1495353 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 17:02:47,301 INFO MainThread:1495353 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 17:02:47,301 INFO MainThread:1495353 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_170247-9rluwifr/logs/debug.log +2025-09-15 17:02:47,302 INFO MainThread:1495353 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_170247-9rluwifr/logs/debug-internal.log +2025-09-15 17:02:47,302 INFO MainThread:1495353 [wandb_init.py:init():813] calling init triggers +2025-09-15 17:02:47,302 INFO MainThread:1495353 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 17:02:47,302 INFO MainThread:1495353 [wandb_init.py:init():854] starting backend +2025-09-15 17:02:47,580 INFO MainThread:1495353 [wandb_init.py:init():857] sending inform_init request +2025-09-15 17:02:47,599 INFO MainThread:1495353 [wandb_init.py:init():865] backend started and connected +2025-09-15 17:02:47,600 INFO MainThread:1495353 [wandb_init.py:init():936] updated telemetry +2025-09-15 17:02:47,607 INFO MainThread:1495353 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 17:02:49,775 INFO MainThread:1495353 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 17:02:50,150 INFO MainThread:1495353 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 17:02:50,151 INFO MainThread:1495353 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 17:02:50,152 INFO MainThread:1495353 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 17:02:50,152 INFO MainThread:1495353 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 17:02:50,183 INFO MainThread:1495353 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 17:02:50,624 INFO wandb-AsyncioManager-main:1495353 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 17:02:50,624 INFO wandb-AsyncioManager-main:1495353 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_170247-9rluwifr/run-9rluwifr.wandb b/wandb/run-20250915_170247-9rluwifr/run-9rluwifr.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250915_170938-mlb6ufl5/files/config.yaml b/wandb/run-20250915_170938-mlb6ufl5/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250915_170938-mlb6ufl5/files/output.log b/wandb/run-20250915_170938-mlb6ufl5/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..61f3b34033a6b870f58508d6dea08fb8bd0da516 --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/files/output.log @@ -0,0 +1,32 @@ +The length of the dataset is: 14409 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 473, numel = 3938312 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 879 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 100%|██████████| 1801/1801 [5:33:55<00:00, 0.09it/s, v_num=ufl5]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_word +Token indices sequence length is longer than the specified maximum sequence length for this model (1028 > 1024). Running this sequence through the model will result in indexing errors +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. +Traceback (most recent call last): + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/serialization.py", line 967, in save + _save( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/serialization.py", line 1268, in _save + zip_file.write_record(name, storage, num_bytes) +RuntimeError: [enforce fail at inline_container.cc:858] . PytorchStreamWriter failed writing file data/1051: file write failed + +During handling of the above exception, another exception occurred: + +Traceback (most recent call last): diff --git a/wandb/run-20250915_170938-mlb6ufl5/files/requirements.txt b/wandb/run-20250915_170938-mlb6ufl5/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_170938-mlb6ufl5/files/wandb-metadata.json b/wandb/run-20250915_170938-mlb6ufl5/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..dc7c5a9bed00d94b5833c378b5a46f7662927dd9 --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T09:09:38.361945Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3578274852864" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "tpxxcyp7ahwk7z6yf7fo0hmarrxrkgel" +} \ No newline at end of file diff --git a/wandb/run-20250915_170938-mlb6ufl5/files/wandb-summary.json b/wandb/run-20250915_170938-mlb6ufl5/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250915_170938-mlb6ufl5/logs/debug-core.log b/wandb/run-20250915_170938-mlb6ufl5/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..fbdcd823349772c943df639d2c51b1c647a2299e --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T17:09:38.397875754+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmprh9clkks/port-1515467.txt","pid":1515467,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T17:09:38.398090359+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T17:09:38.399531355+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":1515467} +{"time":"2025-09-15T17:09:38.399366539+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-1515467-1516217-3738013219/socket","Net":"unix"}} +{"time":"2025-09-15T17:09:38.585511212+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T17:09:38.600084715+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"mlb6ufl5","id":"1(@)"} +{"time":"2025-09-15T17:09:39.066004756+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"mlb6ufl5","id":"1(@)"} +{"time":"2025-09-15T22:45:13.302463352+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T22:45:13.303211626+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T22:45:13.303165554+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T22:45:13.303319814+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T22:45:13.303875836+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-1515467-1516217-3738013219/socket","Net":"unix"}} +{"time":"2025-09-15T22:45:14.312445265+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T22:45:14.312484129+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T22:45:14.312502739+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_170938-mlb6ufl5/logs/debug-internal.log b/wandb/run-20250915_170938-mlb6ufl5/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..04fc95395a99be56b251e8e00d237c543bb55236 --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/logs/debug-internal.log @@ -0,0 +1,14 @@ +{"time":"2025-09-15T17:09:38.600449841+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T17:09:38.625486139+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T17:09:39.065903153+08:00","level":"INFO","msg":"stream: created new stream","id":"mlb6ufl5"} +{"time":"2025-09-15T17:09:39.06598994+08:00","level":"INFO","msg":"stream: started","id":"mlb6ufl5"} +{"time":"2025-09-15T17:09:39.066017236+08:00","level":"INFO","msg":"writer: started","stream_id":"mlb6ufl5"} +{"time":"2025-09-15T17:09:39.066030951+08:00","level":"INFO","msg":"handler: started","stream_id":"mlb6ufl5"} +{"time":"2025-09-15T17:09:39.066066086+08:00","level":"INFO","msg":"sender: started","stream_id":"mlb6ufl5"} +{"time":"2025-09-15T22:45:13.303043193+08:00","level":"INFO","msg":"stream: closing","id":"mlb6ufl5"} +{"time":"2025-09-15T22:45:13.320745853+08:00","level":"ERROR","msg":"sender: failed to upload run summary: write wandb/run-20250915_170938-mlb6ufl5/files/wandb-summary.json: no space left on device"} +{"time":"2025-09-15T22:45:13.322304789+08:00","level":"ERROR","msg":"sender: failed to upload run config: write wandb/run-20250915_170938-mlb6ufl5/files/config.yaml: no space left on device"} +{"time":"2025-09-15T22:45:13.991160105+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T22:45:14.309992685+08:00","level":"INFO","msg":"handler: closed","stream_id":"mlb6ufl5"} +{"time":"2025-09-15T22:45:14.310766581+08:00","level":"INFO","msg":"sender: closed","stream_id":"mlb6ufl5"} +{"time":"2025-09-15T22:45:14.310831549+08:00","level":"INFO","msg":"stream: closed","id":"mlb6ufl5"} diff --git a/wandb/run-20250915_170938-mlb6ufl5/logs/debug.log b/wandb/run-20250915_170938-mlb6ufl5/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7c66eca9a2bf1d7174e518fa01f7ab8909c00c43 --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-15 17:09:38,365 INFO MainThread:1515467 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_setup.py:_flush():81] Configure stats pid to 1515467 +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_170938-mlb6ufl5/logs/debug.log +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_170938-mlb6ufl5/logs/debug-internal.log +2025-09-15 17:09:38,366 INFO MainThread:1515467 [wandb_init.py:init():813] calling init triggers +2025-09-15 17:09:38,367 INFO MainThread:1515467 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 17:09:38,367 INFO MainThread:1515467 [wandb_init.py:init():854] starting backend +2025-09-15 17:09:38,586 INFO MainThread:1515467 [wandb_init.py:init():857] sending inform_init request +2025-09-15 17:09:38,591 INFO MainThread:1515467 [wandb_init.py:init():865] backend started and connected +2025-09-15 17:09:38,595 INFO MainThread:1515467 [wandb_init.py:init():936] updated telemetry +2025-09-15 17:09:38,606 INFO MainThread:1515467 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 17:09:39,330 INFO MainThread:1515467 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 17:09:39,513 INFO MainThread:1515467 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 17:09:39,513 INFO MainThread:1515467 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 17:09:39,513 INFO MainThread:1515467 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 17:09:39,513 INFO MainThread:1515467 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 17:09:39,516 INFO MainThread:1515467 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 17:11:11,980 INFO MainThread:1515467 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 8, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-15 22:45:13,302 INFO wandb-AsyncioManager-main:1515467 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 22:45:13,303 INFO wandb-AsyncioManager-main:1515467 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_170938-mlb6ufl5/run-mlb6ufl5.wandb b/wandb/run-20250915_170938-mlb6ufl5/run-mlb6ufl5.wandb new file mode 100644 index 0000000000000000000000000000000000000000..aaa41f8667cab990d6660516e37a35db90751417 --- /dev/null +++ b/wandb/run-20250915_170938-mlb6ufl5/run-mlb6ufl5.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04937bc8e65fc0e6809334d9a69a03764087fefef4d13065e6c71277565139af +size 1236542 diff --git a/wandb/run-20250915_231258-pq7hjf3o/files/config.yaml b/wandb/run-20250915_231258-pq7hjf3o/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d90420e65d3791479991f3a2668222d989f0db4a --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + ypl6mc1nmmm62ial685wtfhqe5m4tjs5: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=8 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495358074880" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:12:58.266784Z" + writerId: ypl6mc1nmmm62ial685wtfhqe5m4tjs5 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250915_231258-pq7hjf3o/files/output.log b/wandb/run-20250915_231258-pq7hjf3o/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d493245da2fadd32b0fefd5733cf67ddeeb3c32 --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/files/output.log @@ -0,0 +1,56 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 356, in read_data + for message in game["history"]: +TypeError: string indices must be integers +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 356, in read_data +[rank0]: for message in game["history"]: +[rank0]: TypeError: string indices must be integers diff --git a/wandb/run-20250915_231258-pq7hjf3o/files/requirements.txt b/wandb/run-20250915_231258-pq7hjf3o/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_231258-pq7hjf3o/files/wandb-metadata.json b/wandb/run-20250915_231258-pq7hjf3o/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..0a63848d5505a50a4699d9d227344306c52b6ea2 --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:12:58.266784Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495358074880" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ypl6mc1nmmm62ial685wtfhqe5m4tjs5" +} \ No newline at end of file diff --git a/wandb/run-20250915_231258-pq7hjf3o/files/wandb-summary.json b/wandb/run-20250915_231258-pq7hjf3o/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250915_231258-pq7hjf3o/logs/debug-core.log b/wandb/run-20250915_231258-pq7hjf3o/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..ba2ab4ecf7fe36565dcf395f1e4e5f5078e8ec4c --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:12:58.299199536+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp7yh4m3fr/port-2145763.txt","pid":2145763,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:12:58.299467831+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:12:58.300162014+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2145763} +{"time":"2025-09-15T23:12:58.300152171+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2145763-2146352-957498721/socket","Net":"unix"}} +{"time":"2025-09-15T23:12:58.487408147+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:12:58.495671092+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"pq7hjf3o","id":"1(@)"} +{"time":"2025-09-15T23:12:58.996018023+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"pq7hjf3o","id":"1(@)"} +{"time":"2025-09-15T23:12:59.934463208+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T23:12:59.934560009+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T23:12:59.93473401+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T23:12:59.934735343+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T23:12:59.93545156+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2145763-2146352-957498721/socket","Net":"unix"}} +{"time":"2025-09-15T23:13:01.811310275+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T23:13:01.811360324+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T23:13:01.811379738+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_231258-pq7hjf3o/logs/debug-internal.log b/wandb/run-20250915_231258-pq7hjf3o/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e82f88f56efc2199ccae7c88c06f7caf40a0d4c2 --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:12:58.495824457+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:12:58.516621655+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:12:58.995960337+08:00","level":"INFO","msg":"stream: created new stream","id":"pq7hjf3o"} +{"time":"2025-09-15T23:12:58.996010986+08:00","level":"INFO","msg":"stream: started","id":"pq7hjf3o"} +{"time":"2025-09-15T23:12:58.996060727+08:00","level":"INFO","msg":"handler: started","stream_id":"pq7hjf3o"} +{"time":"2025-09-15T23:12:58.996050069+08:00","level":"INFO","msg":"writer: started","stream_id":"pq7hjf3o"} +{"time":"2025-09-15T23:12:58.996065411+08:00","level":"INFO","msg":"sender: started","stream_id":"pq7hjf3o"} +{"time":"2025-09-15T23:12:59.934587106+08:00","level":"INFO","msg":"stream: closing","id":"pq7hjf3o"} +{"time":"2025-09-15T23:13:01.263985355+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T23:13:01.808972631+08:00","level":"INFO","msg":"handler: closed","stream_id":"pq7hjf3o"} +{"time":"2025-09-15T23:13:01.809840266+08:00","level":"INFO","msg":"sender: closed","stream_id":"pq7hjf3o"} +{"time":"2025-09-15T23:13:01.809908369+08:00","level":"INFO","msg":"stream: closed","id":"pq7hjf3o"} diff --git a/wandb/run-20250915_231258-pq7hjf3o/logs/debug.log b/wandb/run-20250915_231258-pq7hjf3o/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..f7657466d3d915a35a2557866b18dbf3230a6dd6 --- /dev/null +++ b/wandb/run-20250915_231258-pq7hjf3o/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 23:12:58,270 INFO MainThread:2145763 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:12:58,270 INFO MainThread:2145763 [wandb_setup.py:_flush():81] Configure stats pid to 2145763 +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_231258-pq7hjf3o/logs/debug.log +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_231258-pq7hjf3o/logs/debug-internal.log +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:12:58,271 INFO MainThread:2145763 [wandb_init.py:init():854] starting backend +2025-09-15 23:12:58,487 INFO MainThread:2145763 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:12:58,493 INFO MainThread:2145763 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:12:58,497 INFO MainThread:2145763 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:12:58,508 INFO MainThread:2145763 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:12:59,373 INFO MainThread:2145763 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:12:59,519 INFO MainThread:2145763 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:12:59,519 INFO MainThread:2145763 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:12:59,519 INFO MainThread:2145763 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:12:59,519 INFO MainThread:2145763 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:12:59,523 INFO MainThread:2145763 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:12:59,935 INFO wandb-AsyncioManager-main:2145763 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 23:12:59,935 INFO wandb-AsyncioManager-main:2145763 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_231258-pq7hjf3o/run-pq7hjf3o.wandb b/wandb/run-20250915_231258-pq7hjf3o/run-pq7hjf3o.wandb new file mode 100644 index 0000000000000000000000000000000000000000..1758f1ff88fcb2215c168e34a0b37a65a311efa0 Binary files /dev/null and b/wandb/run-20250915_231258-pq7hjf3o/run-pq7hjf3o.wandb differ diff --git a/wandb/run-20250915_231323-e2f07hn1/files/config.yaml b/wandb/run-20250915_231323-e2f07hn1/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7d65c973b199a0eb9dcca6fc0463d90dfb54d005 --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + zo6ul3eab3f5o8bc05vgim5dve020983: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495358222336" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:13:23.235167Z" + writerId: zo6ul3eab3f5o8bc05vgim5dve020983 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250915_231323-e2f07hn1/files/output.log b/wandb/run-20250915_231323-e2f07hn1/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d493245da2fadd32b0fefd5733cf67ddeeb3c32 --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/files/output.log @@ -0,0 +1,56 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 356, in read_data + for message in game["history"]: +TypeError: string indices must be integers +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 356, in read_data +[rank0]: for message in game["history"]: +[rank0]: TypeError: string indices must be integers diff --git a/wandb/run-20250915_231323-e2f07hn1/files/requirements.txt b/wandb/run-20250915_231323-e2f07hn1/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_231323-e2f07hn1/files/wandb-metadata.json b/wandb/run-20250915_231323-e2f07hn1/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4b8ba2c6c8e544921a27932b4ee9f09ef0687f9a --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:13:23.235167Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495358222336" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "zo6ul3eab3f5o8bc05vgim5dve020983" +} \ No newline at end of file diff --git a/wandb/run-20250915_231323-e2f07hn1/files/wandb-summary.json b/wandb/run-20250915_231323-e2f07hn1/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b0a620d0c1047a4dd8a400939b6da246ed8063a7 --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0},"_runtime":0} \ No newline at end of file diff --git a/wandb/run-20250915_231323-e2f07hn1/logs/debug-core.log b/wandb/run-20250915_231323-e2f07hn1/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..23a6335fe720f0328ae8d97d70196e4ce4a3f372 --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:13:23.270750717+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpjaqxy46d/port-2147212.txt","pid":2147212,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:13:23.271028959+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:13:23.272026135+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2147212} +{"time":"2025-09-15T23:13:23.272039941+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2147212-2147951-1937249160/socket","Net":"unix"}} +{"time":"2025-09-15T23:13:23.458867503+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:13:23.467569267+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"e2f07hn1","id":"1(@)"} +{"time":"2025-09-15T23:13:23.91953793+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"e2f07hn1","id":"1(@)"} +{"time":"2025-09-15T23:13:24.755282622+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T23:13:24.755675707+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T23:13:24.755712162+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T23:13:24.755768541+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T23:13:24.756034983+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2147212-2147951-1937249160/socket","Net":"unix"}} +{"time":"2025-09-15T23:13:25.916258614+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T23:13:25.916308242+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T23:13:25.916327868+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_231323-e2f07hn1/logs/debug-internal.log b/wandb/run-20250915_231323-e2f07hn1/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..48fe491decb5584c88d9cb978a2d568187a134cd --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:13:23.467746786+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:13:23.482719003+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:13:23.919394579+08:00","level":"INFO","msg":"stream: created new stream","id":"e2f07hn1"} +{"time":"2025-09-15T23:13:23.919522064+08:00","level":"INFO","msg":"stream: started","id":"e2f07hn1"} +{"time":"2025-09-15T23:13:23.919553677+08:00","level":"INFO","msg":"writer: started","stream_id":"e2f07hn1"} +{"time":"2025-09-15T23:13:23.919692798+08:00","level":"INFO","msg":"sender: started","stream_id":"e2f07hn1"} +{"time":"2025-09-15T23:13:23.919736889+08:00","level":"INFO","msg":"handler: started","stream_id":"e2f07hn1"} +{"time":"2025-09-15T23:13:24.755730625+08:00","level":"INFO","msg":"stream: closing","id":"e2f07hn1"} +{"time":"2025-09-15T23:13:25.628301028+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T23:13:25.91455024+08:00","level":"INFO","msg":"handler: closed","stream_id":"e2f07hn1"} +{"time":"2025-09-15T23:13:25.915183225+08:00","level":"INFO","msg":"sender: closed","stream_id":"e2f07hn1"} +{"time":"2025-09-15T23:13:25.915278211+08:00","level":"INFO","msg":"stream: closed","id":"e2f07hn1"} diff --git a/wandb/run-20250915_231323-e2f07hn1/logs/debug.log b/wandb/run-20250915_231323-e2f07hn1/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a62db5d8448ef15185f3fbbc63c1bed7b3277aea --- /dev/null +++ b/wandb/run-20250915_231323-e2f07hn1/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 23:13:23,238 INFO MainThread:2147212 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_setup.py:_flush():81] Configure stats pid to 2147212 +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_231323-e2f07hn1/logs/debug.log +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_231323-e2f07hn1/logs/debug-internal.log +2025-09-15 23:13:23,239 INFO MainThread:2147212 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:13:23,240 INFO MainThread:2147212 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:13:23,240 INFO MainThread:2147212 [wandb_init.py:init():854] starting backend +2025-09-15 23:13:23,459 INFO MainThread:2147212 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:13:23,463 INFO MainThread:2147212 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:13:23,466 INFO MainThread:2147212 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:13:23,477 INFO MainThread:2147212 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:13:24,218 INFO MainThread:2147212 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:13:24,382 INFO MainThread:2147212 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:13:24,382 INFO MainThread:2147212 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:13:24,382 INFO MainThread:2147212 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:13:24,382 INFO MainThread:2147212 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:13:24,386 INFO MainThread:2147212 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:13:24,756 INFO wandb-AsyncioManager-main:2147212 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 23:13:24,756 INFO wandb-AsyncioManager-main:2147212 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_231323-e2f07hn1/run-e2f07hn1.wandb b/wandb/run-20250915_231323-e2f07hn1/run-e2f07hn1.wandb new file mode 100644 index 0000000000000000000000000000000000000000..8bd24b9136a87f8c819460310d97a4b76ccc49a0 Binary files /dev/null and b/wandb/run-20250915_231323-e2f07hn1/run-e2f07hn1.wandb differ diff --git a/wandb/run-20250915_231501-ioat5bkl/files/config.yaml b/wandb/run-20250915_231501-ioat5bkl/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d0b6a2699e6e6ce7aa8dd3fcfb08bf8593a521f1 --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + n752kz2uag3gb2ynkjme9j7gcz1mso1q: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=8 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495358550016" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:15:01.451981Z" + writerId: n752kz2uag3gb2ynkjme9j7gcz1mso1q + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250915_231501-ioat5bkl/files/output.log b/wandb/run-20250915_231501-ioat5bkl/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..419577b6dc03e2abac56ddfa3cb8019b87a19aab --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/files/output.log @@ -0,0 +1,56 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 404, in read_data + "object_list": game["object_list"], +KeyError: 'object_list' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 404, in read_data +[rank0]: "object_list": game["object_list"], +[rank0]: KeyError: 'object_list' diff --git a/wandb/run-20250915_231501-ioat5bkl/files/requirements.txt b/wandb/run-20250915_231501-ioat5bkl/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_231501-ioat5bkl/files/wandb-metadata.json b/wandb/run-20250915_231501-ioat5bkl/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..493072d03d1a04ac04195d7be5ad8847bca99131 --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:15:01.451981Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495358550016" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "n752kz2uag3gb2ynkjme9j7gcz1mso1q" +} \ No newline at end of file diff --git a/wandb/run-20250915_231501-ioat5bkl/files/wandb-summary.json b/wandb/run-20250915_231501-ioat5bkl/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..cce42b8af4ec53c3690f647e729b5f3e4d8e1f41 --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":70,"_wandb":{"runtime":70}} \ No newline at end of file diff --git a/wandb/run-20250915_231501-ioat5bkl/logs/debug-core.log b/wandb/run-20250915_231501-ioat5bkl/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..7eb8ac8a7dd90c404ba9af25c928abcee3f96cfc --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:15:01.487750144+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp64v7t3ng/port-2151074.txt","pid":2151074,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:15:01.488024034+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:15:01.489123771+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2151074} +{"time":"2025-09-15T23:15:01.489162582+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2151074-2151752-4267970548/socket","Net":"unix"}} +{"time":"2025-09-15T23:15:01.677794545+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:15:01.691475935+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ioat5bkl","id":"1(@)"} +{"time":"2025-09-15T23:15:02.152618885+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ioat5bkl","id":"1(@)"} +{"time":"2025-09-15T23:16:12.909803049+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T23:16:12.90992404+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T23:16:12.909902223+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T23:16:12.910055888+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T23:16:12.91022346+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2151074-2151752-4267970548/socket","Net":"unix"}} +{"time":"2025-09-15T23:16:13.883949301+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T23:16:13.884005276+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T23:16:13.884029565+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_231501-ioat5bkl/logs/debug-internal.log b/wandb/run-20250915_231501-ioat5bkl/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..bda1b6267bb9ecd1cd908421fbb383c459690e05 --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:15:01.691617742+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:15:01.712367751+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:15:02.152525734+08:00","level":"INFO","msg":"stream: created new stream","id":"ioat5bkl"} +{"time":"2025-09-15T23:15:02.152603892+08:00","level":"INFO","msg":"stream: started","id":"ioat5bkl"} +{"time":"2025-09-15T23:15:02.1527489+08:00","level":"INFO","msg":"handler: started","stream_id":"ioat5bkl"} +{"time":"2025-09-15T23:15:02.152737458+08:00","level":"INFO","msg":"writer: started","stream_id":"ioat5bkl"} +{"time":"2025-09-15T23:15:02.152746781+08:00","level":"INFO","msg":"sender: started","stream_id":"ioat5bkl"} +{"time":"2025-09-15T23:16:12.909906727+08:00","level":"INFO","msg":"stream: closing","id":"ioat5bkl"} +{"time":"2025-09-15T23:16:13.491979948+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T23:16:13.882837359+08:00","level":"INFO","msg":"handler: closed","stream_id":"ioat5bkl"} +{"time":"2025-09-15T23:16:13.883256222+08:00","level":"INFO","msg":"sender: closed","stream_id":"ioat5bkl"} +{"time":"2025-09-15T23:16:13.88329411+08:00","level":"INFO","msg":"stream: closed","id":"ioat5bkl"} diff --git a/wandb/run-20250915_231501-ioat5bkl/logs/debug.log b/wandb/run-20250915_231501-ioat5bkl/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..ada4bbeb3c6c15171c0b0fc200262ed5b3789d45 --- /dev/null +++ b/wandb/run-20250915_231501-ioat5bkl/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_setup.py:_flush():81] Configure stats pid to 2151074 +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_231501-ioat5bkl/logs/debug.log +2025-09-15 23:15:01,456 INFO MainThread:2151074 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_231501-ioat5bkl/logs/debug-internal.log +2025-09-15 23:15:01,457 INFO MainThread:2151074 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:15:01,457 INFO MainThread:2151074 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:15:01,457 INFO MainThread:2151074 [wandb_init.py:init():854] starting backend +2025-09-15 23:15:01,678 INFO MainThread:2151074 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:15:01,683 INFO MainThread:2151074 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:15:01,687 INFO MainThread:2151074 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:15:01,699 INFO MainThread:2151074 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:15:02,533 INFO MainThread:2151074 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:15:02,711 INFO MainThread:2151074 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:15:02,712 INFO MainThread:2151074 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:15:02,712 INFO MainThread:2151074 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:15:02,712 INFO MainThread:2151074 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:15:02,715 INFO MainThread:2151074 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:16:12,910 INFO wandb-AsyncioManager-main:2151074 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 23:16:12,910 INFO wandb-AsyncioManager-main:2151074 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_231501-ioat5bkl/run-ioat5bkl.wandb b/wandb/run-20250915_231501-ioat5bkl/run-ioat5bkl.wandb new file mode 100644 index 0000000000000000000000000000000000000000..090dc3cafed2524abd13896b5966bcd1eb9ffdb8 Binary files /dev/null and b/wandb/run-20250915_231501-ioat5bkl/run-ioat5bkl.wandb differ diff --git a/wandb/run-20250915_231635-563w7e53/files/config.yaml b/wandb/run-20250915_231635-563w7e53/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..61c35ec69d8e7a89a3a978ced5a0499cd77040d5 --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + 41thxbmuvx5o7ax3yjnvgy5gc2331b6e: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495358795776" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:16:35.974422Z" + writerId: 41thxbmuvx5o7ax3yjnvgy5gc2331b6e + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250915_231635-563w7e53/files/output.log b/wandb/run-20250915_231635-563w7e53/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..419577b6dc03e2abac56ddfa3cb8019b87a19aab --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/files/output.log @@ -0,0 +1,56 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 404, in read_data + "object_list": game["object_list"], +KeyError: 'object_list' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 404, in read_data +[rank0]: "object_list": game["object_list"], +[rank0]: KeyError: 'object_list' diff --git a/wandb/run-20250915_231635-563w7e53/files/requirements.txt b/wandb/run-20250915_231635-563w7e53/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_231635-563w7e53/files/wandb-metadata.json b/wandb/run-20250915_231635-563w7e53/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..f88e3308800a9966f3307eadf7c2819729b644b4 --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:16:35.974422Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495358795776" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "41thxbmuvx5o7ax3yjnvgy5gc2331b6e" +} \ No newline at end of file diff --git a/wandb/run-20250915_231635-563w7e53/files/wandb-summary.json b/wandb/run-20250915_231635-563w7e53/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..9faac867667e86a89d06ff25c1504af87b60de18 --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":47,"_wandb":{"runtime":47}} \ No newline at end of file diff --git a/wandb/run-20250915_231635-563w7e53/logs/debug-core.log b/wandb/run-20250915_231635-563w7e53/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..210655ce4e48ee6f4bf10a425bb28a9207f8f67d --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:16:36.011839827+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmph9bqy1r2/port-2154759.txt","pid":2154759,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:16:36.012043159+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:16:36.012631624+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2154759} +{"time":"2025-09-15T23:16:36.012626793+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2154759-2155468-2885562417/socket","Net":"unix"}} +{"time":"2025-09-15T23:16:36.198539753+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:16:36.207338218+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"563w7e53","id":"1(@)"} +{"time":"2025-09-15T23:16:36.684534996+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"563w7e53","id":"1(@)"} +{"time":"2025-09-15T23:17:24.063845387+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T23:17:24.063950908+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T23:17:24.064113795+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T23:17:24.063967428+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T23:17:24.064364027+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2154759-2155468-2885562417/socket","Net":"unix"}} +{"time":"2025-09-15T23:17:25.110345473+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T23:17:25.110396237+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T23:17:25.110421033+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_231635-563w7e53/logs/debug-internal.log b/wandb/run-20250915_231635-563w7e53/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..218068ff177ad1c57480f9dec5aa98e9120bfa89 --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:16:36.207560577+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:16:36.229894811+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:16:36.684438875+08:00","level":"INFO","msg":"stream: created new stream","id":"563w7e53"} +{"time":"2025-09-15T23:16:36.684521066+08:00","level":"INFO","msg":"stream: started","id":"563w7e53"} +{"time":"2025-09-15T23:16:36.684555596+08:00","level":"INFO","msg":"writer: started","stream_id":"563w7e53"} +{"time":"2025-09-15T23:16:36.684712494+08:00","level":"INFO","msg":"handler: started","stream_id":"563w7e53"} +{"time":"2025-09-15T23:16:36.684739805+08:00","level":"INFO","msg":"sender: started","stream_id":"563w7e53"} +{"time":"2025-09-15T23:17:24.063965491+08:00","level":"INFO","msg":"stream: closing","id":"563w7e53"} +{"time":"2025-09-15T23:17:24.83310404+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T23:17:25.108082346+08:00","level":"INFO","msg":"handler: closed","stream_id":"563w7e53"} +{"time":"2025-09-15T23:17:25.108797164+08:00","level":"INFO","msg":"sender: closed","stream_id":"563w7e53"} +{"time":"2025-09-15T23:17:25.108858705+08:00","level":"INFO","msg":"stream: closed","id":"563w7e53"} diff --git a/wandb/run-20250915_231635-563w7e53/logs/debug.log b/wandb/run-20250915_231635-563w7e53/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..21d172497c1de20bbe6870ab2fc36c7e39dbf293 --- /dev/null +++ b/wandb/run-20250915_231635-563w7e53/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 23:16:35,978 INFO MainThread:2154759 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:16:35,978 INFO MainThread:2154759 [wandb_setup.py:_flush():81] Configure stats pid to 2154759 +2025-09-15 23:16:35,978 INFO MainThread:2154759 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:16:35,978 INFO MainThread:2154759 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:16:35,978 INFO MainThread:2154759 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:16:35,978 INFO MainThread:2154759 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_231635-563w7e53/logs/debug.log +2025-09-15 23:16:35,979 INFO MainThread:2154759 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_231635-563w7e53/logs/debug-internal.log +2025-09-15 23:16:35,979 INFO MainThread:2154759 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:16:35,979 INFO MainThread:2154759 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:16:35,979 INFO MainThread:2154759 [wandb_init.py:init():854] starting backend +2025-09-15 23:16:36,198 INFO MainThread:2154759 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:16:36,203 INFO MainThread:2154759 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:16:36,205 INFO MainThread:2154759 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:16:36,216 INFO MainThread:2154759 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:16:36,953 INFO MainThread:2154759 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:16:37,131 INFO MainThread:2154759 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:16:37,131 INFO MainThread:2154759 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:16:37,132 INFO MainThread:2154759 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:16:37,132 INFO MainThread:2154759 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:16:37,133 INFO MainThread:2154759 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:17:24,064 INFO wandb-AsyncioManager-main:2154759 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 23:17:24,064 INFO wandb-AsyncioManager-main:2154759 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_231635-563w7e53/run-563w7e53.wandb b/wandb/run-20250915_231635-563w7e53/run-563w7e53.wandb new file mode 100644 index 0000000000000000000000000000000000000000..079d0750a8cff32c50fdec197a42ef24d786fecf Binary files /dev/null and b/wandb/run-20250915_231635-563w7e53/run-563w7e53.wandb differ diff --git a/wandb/run-20250915_233828-yn4ninq2/files/config.yaml b/wandb/run-20250915_233828-yn4ninq2/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7260d4f9efb0d055d67c03d2f17b628a9a971da4 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + n2a3gz1iu9l6j5nhbl0om4v0dc18mdls: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=8 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495360286720" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:38:28.667028Z" + writerId: n2a3gz1iu9l6j5nhbl0om4v0dc18mdls + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250915_233828-yn4ninq2/files/output.log b/wandb/run-20250915_233828-yn4ninq2/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..6591adbb334aac27b5ebd49378eaecb65de2cae1 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/files/output.log @@ -0,0 +1,68 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 433, in read_data + outcome, history_length = get_game_outcome( + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 161, in get_game_outcome + if has_target_word(item["content"], target_word): + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 124, in has_target_word + derivative_words = get_derivative_words(target_word) + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 106, in get_derivative_words + word = word.lower() +AttributeError: 'list' object has no attribute 'lower' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 433, in read_data +[rank0]: outcome, history_length = get_game_outcome( +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 161, in get_game_outcome +[rank0]: if has_target_word(item["content"], target_word): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 124, in has_target_word +[rank0]: derivative_words = get_derivative_words(target_word) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 106, in get_derivative_words +[rank0]: word = word.lower() +[rank0]: AttributeError: 'list' object has no attribute 'lower' diff --git a/wandb/run-20250915_233828-yn4ninq2/files/requirements.txt b/wandb/run-20250915_233828-yn4ninq2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_233828-yn4ninq2/files/wandb-metadata.json b/wandb/run-20250915_233828-yn4ninq2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..728568ca4623baf8a31ad0db5245439e7f7b08c6 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:38:28.667028Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495360286720" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "n2a3gz1iu9l6j5nhbl0om4v0dc18mdls" +} \ No newline at end of file diff --git a/wandb/run-20250915_233828-yn4ninq2/files/wandb-summary.json b/wandb/run-20250915_233828-yn4ninq2/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..65efeb016520ebca12e5019029f5a6f6602702bd --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":73},"_runtime":73} \ No newline at end of file diff --git a/wandb/run-20250915_233828-yn4ninq2/logs/debug-core.log b/wandb/run-20250915_233828-yn4ninq2/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..936fcfed94daa1b63b250e80472c2a36c366a681 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:38:28.704928761+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmppg71ts29/port-2194386.txt","pid":2194386,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:38:28.705240203+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:38:28.706116352+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2194386} +{"time":"2025-09-15T23:38:28.706080997+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2194386-2195189-166202228/socket","Net":"unix"}} +{"time":"2025-09-15T23:38:28.892587804+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:38:28.901050156+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"yn4ninq2","id":"1(@)"} +{"time":"2025-09-15T23:38:29.376532414+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"yn4ninq2","id":"1(@)"} +{"time":"2025-09-15T23:39:43.436298099+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T23:39:43.436401727+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T23:39:43.436449493+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T23:39:43.436569708+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T23:39:43.436729972+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2194386-2195189-166202228/socket","Net":"unix"}} +{"time":"2025-09-15T23:39:44.437699206+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T23:39:44.437748873+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T23:39:44.437774138+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_233828-yn4ninq2/logs/debug-internal.log b/wandb/run-20250915_233828-yn4ninq2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..5e5fb86e9a6ef280b803ce2b5b328304fa100ad5 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:38:28.90121189+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:38:28.927306423+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:38:29.376422573+08:00","level":"INFO","msg":"stream: created new stream","id":"yn4ninq2"} +{"time":"2025-09-15T23:38:29.376516414+08:00","level":"INFO","msg":"stream: started","id":"yn4ninq2"} +{"time":"2025-09-15T23:38:29.376532639+08:00","level":"INFO","msg":"writer: started","stream_id":"yn4ninq2"} +{"time":"2025-09-15T23:38:29.376692807+08:00","level":"INFO","msg":"sender: started","stream_id":"yn4ninq2"} +{"time":"2025-09-15T23:38:29.376725033+08:00","level":"INFO","msg":"handler: started","stream_id":"yn4ninq2"} +{"time":"2025-09-15T23:39:43.43642113+08:00","level":"INFO","msg":"stream: closing","id":"yn4ninq2"} +{"time":"2025-09-15T23:39:44.164655477+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T23:39:44.435787574+08:00","level":"INFO","msg":"handler: closed","stream_id":"yn4ninq2"} +{"time":"2025-09-15T23:39:44.436502576+08:00","level":"INFO","msg":"sender: closed","stream_id":"yn4ninq2"} +{"time":"2025-09-15T23:39:44.436555328+08:00","level":"INFO","msg":"stream: closed","id":"yn4ninq2"} diff --git a/wandb/run-20250915_233828-yn4ninq2/logs/debug.log b/wandb/run-20250915_233828-yn4ninq2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..740357ae0f9a84936db002c71b06a2c852bf43f5 --- /dev/null +++ b/wandb/run-20250915_233828-yn4ninq2/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 23:38:28,671 INFO MainThread:2194386 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:38:28,671 INFO MainThread:2194386 [wandb_setup.py:_flush():81] Configure stats pid to 2194386 +2025-09-15 23:38:28,671 INFO MainThread:2194386 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:38:28,671 INFO MainThread:2194386 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:38:28,671 INFO MainThread:2194386 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:38:28,671 INFO MainThread:2194386 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_233828-yn4ninq2/logs/debug.log +2025-09-15 23:38:28,672 INFO MainThread:2194386 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_233828-yn4ninq2/logs/debug-internal.log +2025-09-15 23:38:28,672 INFO MainThread:2194386 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:38:28,672 INFO MainThread:2194386 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:38:28,672 INFO MainThread:2194386 [wandb_init.py:init():854] starting backend +2025-09-15 23:38:28,892 INFO MainThread:2194386 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:38:28,897 INFO MainThread:2194386 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:38:28,902 INFO MainThread:2194386 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:38:28,914 INFO MainThread:2194386 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:38:29,727 INFO MainThread:2194386 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:38:29,912 INFO MainThread:2194386 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:38:29,912 INFO MainThread:2194386 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:38:29,913 INFO MainThread:2194386 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:38:29,913 INFO MainThread:2194386 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:38:29,917 INFO MainThread:2194386 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:39:43,437 INFO wandb-AsyncioManager-main:2194386 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 23:39:43,437 INFO wandb-AsyncioManager-main:2194386 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_233828-yn4ninq2/run-yn4ninq2.wandb b/wandb/run-20250915_233828-yn4ninq2/run-yn4ninq2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..d92a5a2d4c5d605b54512c58acc392315015a823 Binary files /dev/null and b/wandb/run-20250915_233828-yn4ninq2/run-yn4ninq2.wandb differ diff --git a/wandb/run-20250915_234004-yqhrbt2x/files/config.yaml b/wandb/run-20250915_234004-yqhrbt2x/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c789cea85f77fd47bdecf3548a54c4291589bfa8 --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + ifnwefaet6kfviouz9ajhe9ftf1xfyyh: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495360524288" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:40:04.393303Z" + writerId: ifnwefaet6kfviouz9ajhe9ftf1xfyyh + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250915_234004-yqhrbt2x/files/output.log b/wandb/run-20250915_234004-yqhrbt2x/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..6591adbb334aac27b5ebd49378eaecb65de2cae1 --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/files/output.log @@ -0,0 +1,68 @@ +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run + call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook + _call_lightning_datamodule_hook(trainer, "setup", stage=fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook + return fn(*args, **kwargs) + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup + self.dataset = self.read_data() + File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 433, in read_data + outcome, history_length = get_game_outcome( + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 161, in get_game_outcome + if has_target_word(item["content"], target_word): + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 124, in has_target_word + derivative_words = get_derivative_words(target_word) + File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 106, in get_derivative_words + word = word.lower() +AttributeError: 'list' object has no attribute 'lower' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 973, in _run +[rank0]: call._call_setup_hook(self) # allow user to set up LightningModule in accelerator environment +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 108, in _call_setup_hook +[rank0]: _call_lightning_datamodule_hook(trainer, "setup", stage=fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 199, in _call_lightning_datamodule_hook +[rank0]: return fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 332, in setup +[rank0]: self.dataset = self.read_data() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Tasks.py", line 433, in read_data +[rank0]: outcome, history_length = get_game_outcome( +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 161, in get_game_outcome +[rank0]: if has_target_word(item["content"], target_word): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 124, in has_target_word +[rank0]: derivative_words = get_derivative_words(target_word) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/word_taboo.py", line 106, in get_derivative_words +[rank0]: word = word.lower() +[rank0]: AttributeError: 'list' object has no attribute 'lower' diff --git a/wandb/run-20250915_234004-yqhrbt2x/files/requirements.txt b/wandb/run-20250915_234004-yqhrbt2x/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_234004-yqhrbt2x/files/wandb-metadata.json b/wandb/run-20250915_234004-yqhrbt2x/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..27c2df6d19da6332a3fdbbe123840c847f650b8f --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:40:04.393303Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495360524288" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ifnwefaet6kfviouz9ajhe9ftf1xfyyh" +} \ No newline at end of file diff --git a/wandb/run-20250915_234004-yqhrbt2x/files/wandb-summary.json b/wandb/run-20250915_234004-yqhrbt2x/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..6cd1bee6e3ca6dccd47bd974c513f1ce5d596373 --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":46},"_runtime":46} \ No newline at end of file diff --git a/wandb/run-20250915_234004-yqhrbt2x/logs/debug-core.log b/wandb/run-20250915_234004-yqhrbt2x/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bfad34a82b1f2a1cb06cd29b50fa7113cb04f3b9 --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:40:04.428489997+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpypuvesss/port-2197918.txt","pid":2197918,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:40:04.428785424+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:40:04.431043071+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2197918} +{"time":"2025-09-15T23:40:04.431015536+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2197918-2198630-285368521/socket","Net":"unix"}} +{"time":"2025-09-15T23:40:04.617304328+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:40:04.626417182+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"yqhrbt2x","id":"1(@)"} +{"time":"2025-09-15T23:40:05.114089767+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"yqhrbt2x","id":"1(@)"} +{"time":"2025-09-15T23:40:52.309501568+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-15T23:40:52.309660937+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-15T23:40:52.309771217+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-15T23:40:52.309685234+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-15T23:40:52.309914286+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2197918-2198630-285368521/socket","Net":"unix"}} +{"time":"2025-09-15T23:40:53.319367059+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-15T23:40:53.319414579+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-15T23:40:53.319433549+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_234004-yqhrbt2x/logs/debug-internal.log b/wandb/run-20250915_234004-yqhrbt2x/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2b209fc4b671cc8178f0759eb5d4b64e46c128aa --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:40:04.626554392+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:40:04.660350748+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:40:05.113969183+08:00","level":"INFO","msg":"stream: created new stream","id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:05.114072387+08:00","level":"INFO","msg":"stream: started","id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:05.114105978+08:00","level":"INFO","msg":"writer: started","stream_id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:05.114141799+08:00","level":"INFO","msg":"sender: started","stream_id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:05.114180909+08:00","level":"INFO","msg":"handler: started","stream_id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:52.309636662+08:00","level":"INFO","msg":"stream: closing","id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:53.048243131+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-15T23:40:53.316935541+08:00","level":"INFO","msg":"handler: closed","stream_id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:53.317696228+08:00","level":"INFO","msg":"sender: closed","stream_id":"yqhrbt2x"} +{"time":"2025-09-15T23:40:53.317757505+08:00","level":"INFO","msg":"stream: closed","id":"yqhrbt2x"} diff --git a/wandb/run-20250915_234004-yqhrbt2x/logs/debug.log b/wandb/run-20250915_234004-yqhrbt2x/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..53362688abc50089349479e02358b57b784fe2e6 --- /dev/null +++ b/wandb/run-20250915_234004-yqhrbt2x/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-15 23:40:04,397 INFO MainThread:2197918 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:40:04,397 INFO MainThread:2197918 [wandb_setup.py:_flush():81] Configure stats pid to 2197918 +2025-09-15 23:40:04,397 INFO MainThread:2197918 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:40:04,397 INFO MainThread:2197918 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:40:04,397 INFO MainThread:2197918 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:40:04,397 INFO MainThread:2197918 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_234004-yqhrbt2x/logs/debug.log +2025-09-15 23:40:04,398 INFO MainThread:2197918 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_234004-yqhrbt2x/logs/debug-internal.log +2025-09-15 23:40:04,398 INFO MainThread:2197918 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:40:04,398 INFO MainThread:2197918 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:40:04,398 INFO MainThread:2197918 [wandb_init.py:init():854] starting backend +2025-09-15 23:40:04,617 INFO MainThread:2197918 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:40:04,622 INFO MainThread:2197918 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:40:04,624 INFO MainThread:2197918 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:40:04,636 INFO MainThread:2197918 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:40:05,450 INFO MainThread:2197918 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:40:05,615 INFO MainThread:2197918 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:40:05,616 INFO MainThread:2197918 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:40:05,616 INFO MainThread:2197918 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:40:05,616 INFO MainThread:2197918 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:40:05,618 INFO MainThread:2197918 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:40:52,311 INFO wandb-AsyncioManager-main:2197918 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-15 23:40:52,311 INFO wandb-AsyncioManager-main:2197918 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_234004-yqhrbt2x/run-yqhrbt2x.wandb b/wandb/run-20250915_234004-yqhrbt2x/run-yqhrbt2x.wandb new file mode 100644 index 0000000000000000000000000000000000000000..7f4f7d093627a05c7fbcccb47542c14e6beabb18 Binary files /dev/null and b/wandb/run-20250915_234004-yqhrbt2x/run-yqhrbt2x.wandb differ diff --git a/wandb/run-20250915_234149-solrky9k/files/config.yaml b/wandb/run-20250915_234149-solrky9k/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..9f22c2ab7e47a5272075bd50bd9bc506209ed055 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + eotw6ef9o3cp3xo8iv8lmgnojtgytk61: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=8 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495360733184" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T15:41:49.382685Z" + writerId: eotw6ef9o3cp3xo8iv8lmgnojtgytk61 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 8 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250915_234149-solrky9k/files/output.log b/wandb/run-20250915_234149-solrky9k/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..0ac00e3eb8bc9bf898ebe07597c47dfe439c060c --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/files/output.log @@ -0,0 +1,12 @@ +The length of the dataset is: 28653 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 473, numel = 3938312 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 879 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 5%|▍ | 166/3582 [30:27<10:26:48, 0.09it/s, v_num=ky9k] +[rank: 0] Received SIGTERM: 15 + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250915_234149-solrky9k/files/requirements.txt b/wandb/run-20250915_234149-solrky9k/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250915_234149-solrky9k/files/wandb-metadata.json b/wandb/run-20250915_234149-solrky9k/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..674a9e0115c89b6db766b5da429058cf6900d551 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T15:41:49.382685Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495360733184" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "eotw6ef9o3cp3xo8iv8lmgnojtgytk61" +} \ No newline at end of file diff --git a/wandb/run-20250915_234149-solrky9k/files/wandb-summary.json b/wandb/run-20250915_234149-solrky9k/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..44d99a44c1534d6cade5712a2d7bb425dee2f0a9 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/files/wandb-summary.json @@ -0,0 +1 @@ +{"trainer/global_step":149,"actor/factor.min":1.533203125,"epoch":0,"critic/target_q2.min":-0.08935546875,"critic/target_q2.max":-0.0882568359375,"critic/q1.mean":0.45166015625,"actor/factor.max":1.68359375,"critic/target_q1.mean":0.011810302734375,"critic/target_q1.max":0.013153076171875,"actor/loss":11.4140625,"critic/q2.mean":0.5205078125,"critic/q2.loss":0.23291015625,"actor/log_prob.max":-25.8125,"actor/advantages.min":0.4287109375,"actor/advantages.max":0.51953125,"critic/q1.max":0.49853515625,"_runtime":1961,"critic/v1.mean":0.04034423828125,"_step":2,"critic/v1.loss":8.296966552734375e-05,"critic/q1.loss":0.233642578125,"critic/v2.min":-0.023162841796875,"critic/v1.max":0.0423583984375,"actor/factor.mean":1.609375,"critic/target_q1.min":0.01045989990234375,"critic/v2.loss":0.0004493713495321572,"actor/log_prob.mean":-56.03125,"_wandb":{"runtime":1961},"actor/advantages.mean":0.474609375,"critic/q2.max":0.5732421875,"_timestamp":1.757952683920911e+09,"critic/q1.min":0.4052734375,"critic/v2.mean":-0.0218505859375,"actor/log_prob.min":-86.4375,"critic/v2.max":-0.020538330078125,"critic/v1.min":0.03826904296875,"critic/q2.min":0.4677734375,"critic/target_q2.mean":-0.0888671875} \ No newline at end of file diff --git a/wandb/run-20250915_234149-solrky9k/logs/debug-core.log b/wandb/run-20250915_234149-solrky9k/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..99089b224ebbe816de4e8150ba4a7f6422528c37 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-15T23:41:49.419025239+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpljlcve_u/port-2202133.txt","pid":2202133,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-15T23:41:49.419318848+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-15T23:41:49.420589763+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2202133} +{"time":"2025-09-15T23:41:49.420597529+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2202133-2202943-3207280499/socket","Net":"unix"}} +{"time":"2025-09-15T23:41:49.609535467+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-15T23:41:49.625001372+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"solrky9k","id":"1(@)"} +{"time":"2025-09-15T23:41:50.106758651+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"solrky9k","id":"1(@)"} +{"time":"2025-09-16T00:14:31.986505988+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T00:14:31.986651942+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T00:14:31.986633976+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T00:14:31.986866988+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T00:14:31.986917658+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2202133-2202943-3207280499/socket","Net":"unix"}} +{"time":"2025-09-16T00:14:33.744069374+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T00:14:33.74410625+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T00:14:33.744123375+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250915_234149-solrky9k/logs/debug-internal.log b/wandb/run-20250915_234149-solrky9k/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2009f22bd72a0fa6568c265f7013dfa0f69cc414 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-15T23:41:49.625203004+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-15T23:41:49.640769007+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-15T23:41:50.106594138+08:00","level":"INFO","msg":"stream: created new stream","id":"solrky9k"} +{"time":"2025-09-15T23:41:50.106739976+08:00","level":"INFO","msg":"stream: started","id":"solrky9k"} +{"time":"2025-09-15T23:41:50.106863234+08:00","level":"INFO","msg":"writer: started","stream_id":"solrky9k"} +{"time":"2025-09-15T23:41:50.107016233+08:00","level":"INFO","msg":"handler: started","stream_id":"solrky9k"} +{"time":"2025-09-15T23:41:50.107079862+08:00","level":"INFO","msg":"sender: started","stream_id":"solrky9k"} +{"time":"2025-09-16T00:14:31.986628992+08:00","level":"INFO","msg":"stream: closing","id":"solrky9k"} +{"time":"2025-09-16T00:14:33.282442501+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T00:14:33.742578462+08:00","level":"INFO","msg":"handler: closed","stream_id":"solrky9k"} +{"time":"2025-09-16T00:14:33.743282994+08:00","level":"INFO","msg":"sender: closed","stream_id":"solrky9k"} +{"time":"2025-09-16T00:14:33.743336037+08:00","level":"INFO","msg":"stream: closed","id":"solrky9k"} diff --git a/wandb/run-20250915_234149-solrky9k/logs/debug.log b/wandb/run-20250915_234149-solrky9k/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..2aa60820303fba566549e244aa778ab6bc564411 --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-15 23:41:49,386 INFO MainThread:2202133 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_setup.py:_flush():81] Configure stats pid to 2202133 +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250915_234149-solrky9k/logs/debug.log +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250915_234149-solrky9k/logs/debug-internal.log +2025-09-15 23:41:49,387 INFO MainThread:2202133 [wandb_init.py:init():813] calling init triggers +2025-09-15 23:41:49,388 INFO MainThread:2202133 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-15 23:41:49,388 INFO MainThread:2202133 [wandb_init.py:init():854] starting backend +2025-09-15 23:41:49,609 INFO MainThread:2202133 [wandb_init.py:init():857] sending inform_init request +2025-09-15 23:41:49,614 INFO MainThread:2202133 [wandb_init.py:init():865] backend started and connected +2025-09-15 23:41:49,617 INFO MainThread:2202133 [wandb_init.py:init():936] updated telemetry +2025-09-15 23:41:49,629 INFO MainThread:2202133 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-15 23:41:50,417 INFO MainThread:2202133 [wandb_init.py:init():1011] starting run threads in backend +2025-09-15 23:41:50,605 INFO MainThread:2202133 [wandb_run.py:_console_start():2506] atexit reg +2025-09-15 23:41:50,605 INFO MainThread:2202133 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-15 23:41:50,605 INFO MainThread:2202133 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-15 23:41:50,606 INFO MainThread:2202133 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-15 23:41:50,609 INFO MainThread:2202133 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-15 23:43:53,413 INFO MainThread:2202133 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 8, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 00:14:31,986 INFO wandb-AsyncioManager-main:2202133 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 00:14:31,987 INFO wandb-AsyncioManager-main:2202133 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250915_234149-solrky9k/run-solrky9k.wandb b/wandb/run-20250915_234149-solrky9k/run-solrky9k.wandb new file mode 100644 index 0000000000000000000000000000000000000000..c9ce62dfaefafd26a626eba8034f944c6252403e --- /dev/null +++ b/wandb/run-20250915_234149-solrky9k/run-solrky9k.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10bf2e9c4cd14b65c2522641e512cc1cde5981d7a807bfe118760fd58ce469c3 +size 117816 diff --git a/wandb/run-20250916_001517-sdm2bc8f/files/config.yaml b/wandb/run-20250916_001517-sdm2bc8f/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bc35c436249c19230c92e92fb973193129e7abef --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + jebp0frcr1kigcds54x0yttcv71uc4jh: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Meta-Llama-3-8B-Instruct + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=8 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3495363231744" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-15T16:15:17.793091Z" + writerId: jebp0frcr1kigcds54x0yttcv71uc4jh + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 8 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_001517-sdm2bc8f/files/output.log b/wandb/run-20250916_001517-sdm2bc8f/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4111be5fbc17cb02cee8d3fca87f5471026a380f --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/files/output.log @@ -0,0 +1,72 @@ +The length of the dataset is: 28653 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 473, numel = 3938312 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 879 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 45%|████▍ | 800/1791 [3:32:29<4:23:13, 0.06it/s, v_num=bc8f]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py:643: When saving the DeepSpeed Stage 3 checkpoint, each worker will save a shard of the checkpoint within a directory. If a single file is required after training, see https://lightning.ai/docs/pytorch/stable/advanced/model_parallel.html#deepspeed-zero-stage-3-single-file for instructions. +✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Epoch 0: 89%|████████▉ | 1600/1791 [7:01:59<50:22, 0.06it/s, v_num=bc8f]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +Token indices sequence length is longer than the specified maximum sequence length for this model (1078 > 1024). Running this sequence through the model will result in indexing errors +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Epoch 0: 100%|██████████| 1791/1791 [7:51:34<00:00, 0.06it/s, v_num=bc8f]✅ LoRA adapter saved to: checkpoints/archer_Llama3-8B-I_strategic +❌ Error merging LoRA weights: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 638, in merge_and_save_lora + merged_model = self.actor.model.merge_and_unload() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 938, in merge_and_unload + return self._unload_and_optionally_merge( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 555, in _unload_and_optionally_merge + target.merge(safe_merge=safe_merge, adapter_names=adapter_names) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/layer.py", line 679, in merge + base_layer.weight.data += delta_weight +RuntimeError: The size of tensor a (0) must match the size of tensor b (4096) at non-singleton dimension 1 +`Trainer.fit` stopped: `max_epochs=1` reached. +Epoch 0: 100%|██████████| 1791/1791 [7:51:41<00:00, 0.06it/s, v_num=bc8f] diff --git a/wandb/run-20250916_001517-sdm2bc8f/files/requirements.txt b/wandb/run-20250916_001517-sdm2bc8f/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_001517-sdm2bc8f/files/wandb-metadata.json b/wandb/run-20250916_001517-sdm2bc8f/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7cb8da244a5e378b26a20754567841c84479249a --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-15T16:15:17.793091Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Meta-Llama-3-8B-Instruct", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=8", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Llama3-8B-I_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3495363231744" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "jebp0frcr1kigcds54x0yttcv71uc4jh" +} \ No newline at end of file diff --git a/wandb/run-20250916_001517-sdm2bc8f/files/wandb-summary.json b/wandb/run-20250916_001517-sdm2bc8f/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8aea6bc86cbaa30dd570eade93f605bacd3c3e5d --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/files/wandb-summary.json @@ -0,0 +1 @@ +{"critic/v2.mean":-0.07373046875,"critic/q2.max":1.00244140625,"actor/log_prob.max":-20.6171875,"critic/q1.mean":0.7744140625,"epoch":0,"actor/factor.min":2.119140625,"critic/target_q1.max":0.01314544677734375,"critic/v2.loss":2.1965803171042353e-05,"actor/advantages.mean":0.8486328125,"actor/advantages.min":0.69287109375,"actor/factor.mean":2.4375,"critic/q1.loss":0.06694507598876953,"critic/v1.min":0.0163116455078125,"critic/target_q1.mean":0.01212310791015625,"actor/log_prob.mean":-46.703125,"critic/q2.mean":0.84521484375,"critic/v2.max":-0.07086181640625,"actor/factor.max":2.75390625,"critic/target_q1.min":0.0110931396484375,"critic/v2.min":-0.07666015625,"critic/q1.min":0.6168212890625,"critic/v1.loss":8.848309335007798e-06,"actor/advantages.max":1.00341796875,"_runtime":28405,"critic/q2.min":0.6885986328125,"critic/target_q2.max":-0.08624267578125,"critic/target_q2.mean":-0.08795166015625,"_timestamp":1.7579806822595842e+09,"critic/v1.mean":0.0191192626953125,"actor/loss":14.921875,"actor/log_prob.min":-72.78125,"_step":34,"critic/q2.loss":0.0652322769165039,"critic/target_q2.min":-0.08953857421875,"critic/q1.max":0.93310546875,"trainer/global_step":1749,"critic/v1.max":0.02191162109375,"_wandb":{"runtime":28405}} \ No newline at end of file diff --git a/wandb/run-20250916_001517-sdm2bc8f/logs/debug-core.log b/wandb/run-20250916_001517-sdm2bc8f/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..fc6b208283cfce129fdd0d5bfc15020f72bdb3c7 --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T00:15:17.826191802+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmptah9rjw7/port-2267275.txt","pid":2267275,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T00:15:17.826400811+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T00:15:17.826906497+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2267275} +{"time":"2025-09-16T00:15:17.826898023+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2267275-2268488-1823389132/socket","Net":"unix"}} +{"time":"2025-09-16T00:15:18.015469652+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T00:15:18.024320155+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"sdm2bc8f","id":"1(@)"} +{"time":"2025-09-16T00:15:18.498601315+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"sdm2bc8f","id":"1(@)"} +{"time":"2025-09-16T08:08:44.020311338+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T08:08:44.020467198+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T08:08:44.020446352+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T08:08:44.020603849+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T08:08:44.020935633+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-2267275-2268488-1823389132/socket","Net":"unix"}} +{"time":"2025-09-16T08:08:45.625924977+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T08:08:45.625969286+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T08:08:45.625990277+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_001517-sdm2bc8f/logs/debug-internal.log b/wandb/run-20250916_001517-sdm2bc8f/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8760d1149bdac108b4704b59aee8ee5ac3f43da3 --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T00:15:18.02448838+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T00:15:18.039733006+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T00:15:18.498524803+08:00","level":"INFO","msg":"stream: created new stream","id":"sdm2bc8f"} +{"time":"2025-09-16T00:15:18.498595128+08:00","level":"INFO","msg":"stream: started","id":"sdm2bc8f"} +{"time":"2025-09-16T00:15:18.498768559+08:00","level":"INFO","msg":"writer: started","stream_id":"sdm2bc8f"} +{"time":"2025-09-16T00:15:18.498872343+08:00","level":"INFO","msg":"handler: started","stream_id":"sdm2bc8f"} +{"time":"2025-09-16T00:15:18.498969308+08:00","level":"INFO","msg":"sender: started","stream_id":"sdm2bc8f"} +{"time":"2025-09-16T08:08:44.020426284+08:00","level":"INFO","msg":"stream: closing","id":"sdm2bc8f"} +{"time":"2025-09-16T08:08:45.365136884+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T08:08:45.624661127+08:00","level":"INFO","msg":"handler: closed","stream_id":"sdm2bc8f"} +{"time":"2025-09-16T08:08:45.62521759+08:00","level":"INFO","msg":"sender: closed","stream_id":"sdm2bc8f"} +{"time":"2025-09-16T08:08:45.625277824+08:00","level":"INFO","msg":"stream: closed","id":"sdm2bc8f"} diff --git a/wandb/run-20250916_001517-sdm2bc8f/logs/debug.log b/wandb/run-20250916_001517-sdm2bc8f/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..9667b8e692814aff19af135b40b178e46f01215f --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_setup.py:_flush():81] Configure stats pid to 2267275 +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_001517-sdm2bc8f/logs/debug.log +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_001517-sdm2bc8f/logs/debug-internal.log +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_init.py:init():813] calling init triggers +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 00:15:17,797 INFO MainThread:2267275 [wandb_init.py:init():854] starting backend +2025-09-16 00:15:18,015 INFO MainThread:2267275 [wandb_init.py:init():857] sending inform_init request +2025-09-16 00:15:18,018 INFO MainThread:2267275 [wandb_init.py:init():865] backend started and connected +2025-09-16 00:15:18,020 INFO MainThread:2267275 [wandb_init.py:init():936] updated telemetry +2025-09-16 00:15:18,027 INFO MainThread:2267275 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 00:15:18,811 INFO MainThread:2267275 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 00:15:18,980 INFO MainThread:2267275 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 00:15:18,980 INFO MainThread:2267275 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 00:15:18,981 INFO MainThread:2267275 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 00:15:18,981 INFO MainThread:2267275 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 00:15:18,984 INFO MainThread:2267275 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 00:17:01,414 INFO MainThread:2267275 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Llama3-8B-I_strategic/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 8, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 08:08:44,020 INFO wandb-AsyncioManager-main:2267275 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 08:08:44,020 INFO wandb-AsyncioManager-main:2267275 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_001517-sdm2bc8f/run-sdm2bc8f.wandb b/wandb/run-20250916_001517-sdm2bc8f/run-sdm2bc8f.wandb new file mode 100644 index 0000000000000000000000000000000000000000..db2ad6258b9684ecb2af42393cfb6511e39f054e --- /dev/null +++ b/wandb/run-20250916_001517-sdm2bc8f/run-sdm2bc8f.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b9b3ce5b59597d7773d094d2fda50b49ab4c79b07290ac8015e55e2c2abb992 +size 1494785 diff --git a/wandb/run-20250916_081049-sbqxofmk/files/output.log b/wandb/run-20250916_081049-sbqxofmk/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b255b54de6273667501ee12153deb9a2171c4aa7 --- /dev/null +++ b/wandb/run-20250916_081049-sbqxofmk/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 4/1717 [01:55<13:46:35, 0.03it/s, v_num=ofmk] diff --git a/wandb/run-20250916_081049-sbqxofmk/files/requirements.txt b/wandb/run-20250916_081049-sbqxofmk/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_081049-sbqxofmk/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_081049-sbqxofmk/files/wandb-metadata.json b/wandb/run-20250916_081049-sbqxofmk/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4beef290c7b967a91e183cc78ffa98af4115543e --- /dev/null +++ b/wandb/run-20250916_081049-sbqxofmk/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T00:10:49.404038Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3533864714240" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "jvz4c9a6oc428tr6v4l4z6vovnjqpv5v" +} \ No newline at end of file diff --git a/wandb/run-20250916_081049-sbqxofmk/logs/debug-core.log b/wandb/run-20250916_081049-sbqxofmk/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..1418d23a9c9748298658ddf533f133abde2805e4 --- /dev/null +++ b/wandb/run-20250916_081049-sbqxofmk/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T08:10:49.52710215+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpzxxkq6xx/port-2966209.txt","pid":2966209,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T08:10:49.52744935+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T08:10:49.529679435+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-2966209-2966830-1621610906/socket","Net":"unix"}} +{"time":"2025-09-16T08:10:49.529916243+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":2966209} +{"time":"2025-09-16T08:10:49.709161199+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T08:10:49.74818202+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"sbqxofmk","id":"1(@)"} +{"time":"2025-09-16T08:10:51.073470211+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"sbqxofmk","id":"1(@)"} +{"time":"2025-09-16T08:15:42.89813027+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_081049-sbqxofmk/logs/debug-internal.log b/wandb/run-20250916_081049-sbqxofmk/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..15073020371880391e10896666e37521d77477b8 --- /dev/null +++ b/wandb/run-20250916_081049-sbqxofmk/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T08:10:49.748496236+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T08:10:49.778858527+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T08:10:51.073347657+08:00","level":"INFO","msg":"stream: created new stream","id":"sbqxofmk"} +{"time":"2025-09-16T08:10:51.073448388+08:00","level":"INFO","msg":"stream: started","id":"sbqxofmk"} +{"time":"2025-09-16T08:10:51.073580142+08:00","level":"INFO","msg":"writer: started","stream_id":"sbqxofmk"} +{"time":"2025-09-16T08:10:51.073585766+08:00","level":"INFO","msg":"handler: started","stream_id":"sbqxofmk"} +{"time":"2025-09-16T08:10:51.07372837+08:00","level":"INFO","msg":"sender: started","stream_id":"sbqxofmk"} diff --git a/wandb/run-20250916_081049-sbqxofmk/logs/debug.log b/wandb/run-20250916_081049-sbqxofmk/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6b682ceae80c1c4230f04e3da3b1ffe087df1d50 --- /dev/null +++ b/wandb/run-20250916_081049-sbqxofmk/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 08:10:49,409 INFO MainThread:2966209 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 08:10:49,409 INFO MainThread:2966209 [wandb_setup.py:_flush():81] Configure stats pid to 2966209 +2025-09-16 08:10:49,409 INFO MainThread:2966209 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 08:10:49,409 INFO MainThread:2966209 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 08:10:49,409 INFO MainThread:2966209 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 08:10:49,409 INFO MainThread:2966209 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_081049-sbqxofmk/logs/debug.log +2025-09-16 08:10:49,410 INFO MainThread:2966209 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_081049-sbqxofmk/logs/debug-internal.log +2025-09-16 08:10:49,410 INFO MainThread:2966209 [wandb_init.py:init():813] calling init triggers +2025-09-16 08:10:49,410 INFO MainThread:2966209 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 08:10:49,410 INFO MainThread:2966209 [wandb_init.py:init():854] starting backend +2025-09-16 08:10:49,709 INFO MainThread:2966209 [wandb_init.py:init():857] sending inform_init request +2025-09-16 08:10:49,736 INFO MainThread:2966209 [wandb_init.py:init():865] backend started and connected +2025-09-16 08:10:49,739 INFO MainThread:2966209 [wandb_init.py:init():936] updated telemetry +2025-09-16 08:10:49,752 INFO MainThread:2966209 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 08:10:51,365 INFO MainThread:2966209 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 08:10:51,727 INFO MainThread:2966209 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 08:10:51,727 INFO MainThread:2966209 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 08:10:51,728 INFO MainThread:2966209 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 08:10:51,728 INFO MainThread:2966209 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 08:10:51,761 INFO MainThread:2966209 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 08:12:59,012 INFO MainThread:2966209 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_081049-sbqxofmk/run-sbqxofmk.wandb b/wandb/run-20250916_081049-sbqxofmk/run-sbqxofmk.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_144738-fj4q0mj3/files/config.yaml b/wandb/run-20250916_144738-fj4q0mj3/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e87b5efeeb932a7087c85adefb30a71701c4e876 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + fjhfst5tjolvixnsnmw6fkt1d7ybj7ti: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3559455662080" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T06:47:38.774115Z" + writerId: fjhfst5tjolvixnsnmw6fkt1d7ybj7ti + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_144738-fj4q0mj3/files/output.log b/wandb/run-20250916_144738-fj4q0mj3/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_144738-fj4q0mj3/files/requirements.txt b/wandb/run-20250916_144738-fj4q0mj3/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_144738-fj4q0mj3/files/wandb-metadata.json b/wandb/run-20250916_144738-fj4q0mj3/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e668d983572ee8ce5d87dba75fc856f9df750520 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T06:47:38.774115Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3559455662080" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "fjhfst5tjolvixnsnmw6fkt1d7ybj7ti" +} \ No newline at end of file diff --git a/wandb/run-20250916_144738-fj4q0mj3/files/wandb-summary.json b/wandb/run-20250916_144738-fj4q0mj3/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..1f6a8d6fba6992723d0f745edbf01bee18eb2022 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":27},"_runtime":27} \ No newline at end of file diff --git a/wandb/run-20250916_144738-fj4q0mj3/logs/debug-core.log b/wandb/run-20250916_144738-fj4q0mj3/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..ffd6b213c68d71345aa47ef8339560b477d0e550 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-16T14:47:38.810288974+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpqqpj1r94/port-3174705.txt","pid":3174705,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T14:47:38.810497448+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T14:47:38.811104158+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3174705} +{"time":"2025-09-16T14:47:38.811100253+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3174705-3175868-1700008184/socket","Net":"unix"}} +{"time":"2025-09-16T14:47:38.99983946+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T14:47:39.010152998+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"fj4q0mj3","id":"1(@)"} +{"time":"2025-09-16T14:47:39.5075871+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"fj4q0mj3","id":"1(@)"} +{"time":"2025-09-16T14:48:07.841086124+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T14:48:07.841265966+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T14:48:07.841257722+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T14:48:07.841446112+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T14:48:07.841603237+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3174705-3175868-1700008184/socket","Net":"unix"}} +{"time":"2025-09-16T14:48:08.496954907+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_144738-fj4q0mj3/logs/debug-internal.log b/wandb/run-20250916_144738-fj4q0mj3/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0777ed88c21518e7e11db9a16a57a0ae1e8d6a7d --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T14:47:39.010295071+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T14:47:39.02395171+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T14:47:39.507438538+08:00","level":"INFO","msg":"stream: created new stream","id":"fj4q0mj3"} +{"time":"2025-09-16T14:47:39.507571158+08:00","level":"INFO","msg":"stream: started","id":"fj4q0mj3"} +{"time":"2025-09-16T14:47:39.507586963+08:00","level":"INFO","msg":"writer: started","stream_id":"fj4q0mj3"} +{"time":"2025-09-16T14:47:39.507754108+08:00","level":"INFO","msg":"sender: started","stream_id":"fj4q0mj3"} +{"time":"2025-09-16T14:47:39.507851734+08:00","level":"INFO","msg":"handler: started","stream_id":"fj4q0mj3"} +{"time":"2025-09-16T14:48:07.841203998+08:00","level":"INFO","msg":"stream: closing","id":"fj4q0mj3"} diff --git a/wandb/run-20250916_144738-fj4q0mj3/logs/debug.log b/wandb/run-20250916_144738-fj4q0mj3/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..5828b67ba5d3841c1744e6aa1c5f24b2bd3f1062 --- /dev/null +++ b/wandb/run-20250916_144738-fj4q0mj3/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 14:47:38,778 INFO MainThread:3174705 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 14:47:38,778 INFO MainThread:3174705 [wandb_setup.py:_flush():81] Configure stats pid to 3174705 +2025-09-16 14:47:38,778 INFO MainThread:3174705 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 14:47:38,778 INFO MainThread:3174705 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 14:47:38,778 INFO MainThread:3174705 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 14:47:38,778 INFO MainThread:3174705 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_144738-fj4q0mj3/logs/debug.log +2025-09-16 14:47:38,779 INFO MainThread:3174705 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_144738-fj4q0mj3/logs/debug-internal.log +2025-09-16 14:47:38,779 INFO MainThread:3174705 [wandb_init.py:init():813] calling init triggers +2025-09-16 14:47:38,779 INFO MainThread:3174705 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 14:47:38,779 INFO MainThread:3174705 [wandb_init.py:init():854] starting backend +2025-09-16 14:47:38,999 INFO MainThread:3174705 [wandb_init.py:init():857] sending inform_init request +2025-09-16 14:47:39,005 INFO MainThread:3174705 [wandb_init.py:init():865] backend started and connected +2025-09-16 14:47:39,007 INFO MainThread:3174705 [wandb_init.py:init():936] updated telemetry +2025-09-16 14:47:39,018 INFO MainThread:3174705 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 14:47:39,843 INFO MainThread:3174705 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 14:47:40,028 INFO MainThread:3174705 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 14:47:40,029 INFO MainThread:3174705 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 14:47:40,029 INFO MainThread:3174705 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 14:47:40,029 INFO MainThread:3174705 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 14:47:40,031 INFO MainThread:3174705 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 14:48:07,841 INFO wandb-AsyncioManager-main:3174705 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 14:48:07,841 INFO wandb-AsyncioManager-main:3174705 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_144738-fj4q0mj3/run-fj4q0mj3.wandb b/wandb/run-20250916_144738-fj4q0mj3/run-fj4q0mj3.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_144926-lnqz314h/files/output.log b/wandb/run-20250916_144926-lnqz314h/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..f6af189163d52030f5071fb36393eec9f6fc613f --- /dev/null +++ b/wandb/run-20250916_144926-lnqz314h/files/output.log @@ -0,0 +1,8 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_144926-lnqz314h/files/requirements.txt b/wandb/run-20250916_144926-lnqz314h/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_144926-lnqz314h/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_144926-lnqz314h/files/wandb-metadata.json b/wandb/run-20250916_144926-lnqz314h/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3d6dd64effae9f9add773bb73fbeaec078fb48b9 --- /dev/null +++ b/wandb/run-20250916_144926-lnqz314h/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T06:49:26.730904Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3566771720192" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "bzm4og65ovvd9twfytkc7bjyycorg8b0" +} \ No newline at end of file diff --git a/wandb/run-20250916_144926-lnqz314h/logs/debug-core.log b/wandb/run-20250916_144926-lnqz314h/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..3dfa4d52e112f4365f304be62c2d5aa3f2caa41d --- /dev/null +++ b/wandb/run-20250916_144926-lnqz314h/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T14:49:26.758923655+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpj73khhhx/port-3179484.txt","pid":3179484,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T14:49:26.759137529+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T14:49:26.759807456+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3179484} +{"time":"2025-09-16T14:49:26.75979797+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3179484-3180877-2409053786/socket","Net":"unix"}} +{"time":"2025-09-16T14:49:26.947851963+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T14:49:26.956708876+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"lnqz314h","id":"1(@)"} +{"time":"2025-09-16T14:49:27.464039551+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"lnqz314h","id":"1(@)"} +{"time":"2025-09-16T14:50:32.550076329+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_144926-lnqz314h/logs/debug-internal.log b/wandb/run-20250916_144926-lnqz314h/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..66514519ac7e6aa99c5898ca977ae5288a3e4484 --- /dev/null +++ b/wandb/run-20250916_144926-lnqz314h/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T14:49:26.956869043+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T14:49:26.974117255+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T14:49:27.463859677+08:00","level":"INFO","msg":"stream: created new stream","id":"lnqz314h"} +{"time":"2025-09-16T14:49:27.464020214+08:00","level":"INFO","msg":"stream: started","id":"lnqz314h"} +{"time":"2025-09-16T14:49:27.46417947+08:00","level":"INFO","msg":"writer: started","stream_id":"lnqz314h"} +{"time":"2025-09-16T14:49:27.464362147+08:00","level":"INFO","msg":"handler: started","stream_id":"lnqz314h"} +{"time":"2025-09-16T14:49:27.464377382+08:00","level":"INFO","msg":"sender: started","stream_id":"lnqz314h"} diff --git a/wandb/run-20250916_144926-lnqz314h/logs/debug.log b/wandb/run-20250916_144926-lnqz314h/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..56f46d801a8804dbe8c24a0a357c13f066bf66b5 --- /dev/null +++ b/wandb/run-20250916_144926-lnqz314h/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_setup.py:_flush():81] Configure stats pid to 3179484 +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_144926-lnqz314h/logs/debug.log +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_144926-lnqz314h/logs/debug-internal.log +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_init.py:init():813] calling init triggers +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 14:49:26,734 INFO MainThread:3179484 [wandb_init.py:init():854] starting backend +2025-09-16 14:49:26,948 INFO MainThread:3179484 [wandb_init.py:init():857] sending inform_init request +2025-09-16 14:49:26,952 INFO MainThread:3179484 [wandb_init.py:init():865] backend started and connected +2025-09-16 14:49:26,955 INFO MainThread:3179484 [wandb_init.py:init():936] updated telemetry +2025-09-16 14:49:26,965 INFO MainThread:3179484 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 14:49:27,723 INFO MainThread:3179484 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 14:49:27,892 INFO MainThread:3179484 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 14:49:27,892 INFO MainThread:3179484 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 14:49:27,893 INFO MainThread:3179484 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 14:49:27,893 INFO MainThread:3179484 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 14:49:27,896 INFO MainThread:3179484 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250916_144926-lnqz314h/run-lnqz314h.wandb b/wandb/run-20250916_144926-lnqz314h/run-lnqz314h.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_145113-zvkraf50/files/output.log b/wandb/run-20250916_145113-zvkraf50/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..314ef19c83f9bd75e5d4319f1383498fb3800c1c --- /dev/null +++ b/wandb/run-20250916_145113-zvkraf50/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 4/1717 [02:07<15:12:49, 0.03it/s, v_num=af50] diff --git a/wandb/run-20250916_145113-zvkraf50/files/requirements.txt b/wandb/run-20250916_145113-zvkraf50/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_145113-zvkraf50/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_145113-zvkraf50/files/wandb-metadata.json b/wandb/run-20250916_145113-zvkraf50/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ad1bd5fa57e9325db17533760959abcc3dc9d737 --- /dev/null +++ b/wandb/run-20250916_145113-zvkraf50/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T06:51:13.857172Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3574669279232" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "hgze49pluura21egm2tt22j83qnsvxak" +} \ No newline at end of file diff --git a/wandb/run-20250916_145113-zvkraf50/logs/debug-core.log b/wandb/run-20250916_145113-zvkraf50/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..36e5258955e55b207190673ba88624032bc02266 --- /dev/null +++ b/wandb/run-20250916_145113-zvkraf50/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T14:51:13.89686082+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp9nljipue/port-3183568.txt","pid":3183568,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T14:51:13.897066501+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T14:51:13.897664236+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3183568} +{"time":"2025-09-16T14:51:13.897646442+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3183568-3184264-2162995920/socket","Net":"unix"}} +{"time":"2025-09-16T14:51:14.082422631+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T14:51:14.090950702+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"zvkraf50","id":"1(@)"} +{"time":"2025-09-16T14:51:14.575322957+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"zvkraf50","id":"1(@)"} +{"time":"2025-09-16T14:55:31.312105562+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_145113-zvkraf50/logs/debug-internal.log b/wandb/run-20250916_145113-zvkraf50/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..1098225c665054e0ed48c1e8ca383bcd75746025 --- /dev/null +++ b/wandb/run-20250916_145113-zvkraf50/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T14:51:14.091244965+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T14:51:14.112988947+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T14:51:14.575168781+08:00","level":"INFO","msg":"stream: created new stream","id":"zvkraf50"} +{"time":"2025-09-16T14:51:14.575307104+08:00","level":"INFO","msg":"stream: started","id":"zvkraf50"} +{"time":"2025-09-16T14:51:14.575349203+08:00","level":"INFO","msg":"writer: started","stream_id":"zvkraf50"} +{"time":"2025-09-16T14:51:14.575443926+08:00","level":"INFO","msg":"handler: started","stream_id":"zvkraf50"} +{"time":"2025-09-16T14:51:14.57549484+08:00","level":"INFO","msg":"sender: started","stream_id":"zvkraf50"} diff --git a/wandb/run-20250916_145113-zvkraf50/logs/debug.log b/wandb/run-20250916_145113-zvkraf50/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..5b9a52a15ebbe490ef999c7163a635af9bb9a35f --- /dev/null +++ b/wandb/run-20250916_145113-zvkraf50/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 14:51:13,862 INFO MainThread:3183568 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_setup.py:_flush():81] Configure stats pid to 3183568 +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_145113-zvkraf50/logs/debug.log +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_145113-zvkraf50/logs/debug-internal.log +2025-09-16 14:51:13,863 INFO MainThread:3183568 [wandb_init.py:init():813] calling init triggers +2025-09-16 14:51:13,864 INFO MainThread:3183568 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 14:51:13,864 INFO MainThread:3183568 [wandb_init.py:init():854] starting backend +2025-09-16 14:51:14,082 INFO MainThread:3183568 [wandb_init.py:init():857] sending inform_init request +2025-09-16 14:51:14,085 INFO MainThread:3183568 [wandb_init.py:init():865] backend started and connected +2025-09-16 14:51:14,086 INFO MainThread:3183568 [wandb_init.py:init():936] updated telemetry +2025-09-16 14:51:14,094 INFO MainThread:3183568 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 14:51:14,904 INFO MainThread:3183568 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 14:51:15,088 INFO MainThread:3183568 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 14:51:15,088 INFO MainThread:3183568 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 14:51:15,088 INFO MainThread:3183568 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 14:51:15,088 INFO MainThread:3183568 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 14:51:15,091 INFO MainThread:3183568 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 14:52:33,363 INFO MainThread:3183568 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_145113-zvkraf50/run-zvkraf50.wandb b/wandb/run-20250916_145113-zvkraf50/run-zvkraf50.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_150112-al2s20a2/files/config.yaml b/wandb/run-20250916_150112-al2s20a2/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..591bdefadbf8099948cf51b83a648240b64977eb --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + ez5294btclpmk3cr3ymaty2uihambidv: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577045655552" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:01:12.346658Z" + writerId: ez5294btclpmk3cr3ymaty2uihambidv + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_150112-al2s20a2/files/output.log b/wandb/run-20250916_150112-al2s20a2/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..9f35a4489dfc5d7b7eb920aefa4794480c04908b --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/files/output.log @@ -0,0 +1,12 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 1/826 [00:31<7:11:31, 0.03it/s, v_num=20a2] +[rank: 0] Received SIGTERM: 15 + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_150112-al2s20a2/files/requirements.txt b/wandb/run-20250916_150112-al2s20a2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_150112-al2s20a2/files/wandb-metadata.json b/wandb/run-20250916_150112-al2s20a2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8216a797471295121a25851a74fdae09a2f001b6 --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:01:12.346658Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577045655552" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ez5294btclpmk3cr3ymaty2uihambidv" +} \ No newline at end of file diff --git a/wandb/run-20250916_150112-al2s20a2/files/wandb-summary.json b/wandb/run-20250916_150112-al2s20a2/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..ca39cff3a80ac61d6c83bc5f9e8531cee5288100 --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":147},"_runtime":147} \ No newline at end of file diff --git a/wandb/run-20250916_150112-al2s20a2/logs/debug-core.log b/wandb/run-20250916_150112-al2s20a2/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..baac44056587a2bdc9a6a3072ef09e5815326488 --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:01:12.383668069+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpsixj9s2p/port-3201905.txt","pid":3201905,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:01:12.383866627+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:01:12.384652203+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3201905} +{"time":"2025-09-16T15:01:12.384644803+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3201905-3203260-434776010/socket","Net":"unix"}} +{"time":"2025-09-16T15:01:12.57262136+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:01:12.581339807+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"al2s20a2","id":"1(@)"} +{"time":"2025-09-16T15:01:13.062396263+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"al2s20a2","id":"1(@)"} +{"time":"2025-09-16T15:03:41.227287317+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:03:41.227391884+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:03:41.227389758+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:03:41.227547205+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3201905-3203260-434776010/socket","Net":"unix"}} +{"time":"2025-09-16T15:03:41.227573606+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:03:42.176745095+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:03:42.176796276+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:03:42.176816305+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_150112-al2s20a2/logs/debug-internal.log b/wandb/run-20250916_150112-al2s20a2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d4f324f0f676c2930c425bb896e70d0db1c4d309 --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:01:12.581509979+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:01:12.608260916+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:01:13.062273357+08:00","level":"INFO","msg":"stream: created new stream","id":"al2s20a2"} +{"time":"2025-09-16T15:01:13.062382013+08:00","level":"INFO","msg":"stream: started","id":"al2s20a2"} +{"time":"2025-09-16T15:01:13.062410643+08:00","level":"INFO","msg":"writer: started","stream_id":"al2s20a2"} +{"time":"2025-09-16T15:01:13.062424513+08:00","level":"INFO","msg":"sender: started","stream_id":"al2s20a2"} +{"time":"2025-09-16T15:01:13.062571771+08:00","level":"INFO","msg":"handler: started","stream_id":"al2s20a2"} +{"time":"2025-09-16T15:03:41.227355538+08:00","level":"INFO","msg":"stream: closing","id":"al2s20a2"} +{"time":"2025-09-16T15:03:41.925790844+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:03:42.174362085+08:00","level":"INFO","msg":"handler: closed","stream_id":"al2s20a2"} +{"time":"2025-09-16T15:03:42.17503204+08:00","level":"INFO","msg":"sender: closed","stream_id":"al2s20a2"} +{"time":"2025-09-16T15:03:42.175084289+08:00","level":"INFO","msg":"stream: closed","id":"al2s20a2"} diff --git a/wandb/run-20250916_150112-al2s20a2/logs/debug.log b/wandb/run-20250916_150112-al2s20a2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..594effa704812b6bcfa5f087579d2a9767e9aafe --- /dev/null +++ b/wandb/run-20250916_150112-al2s20a2/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:01:12,350 INFO MainThread:3201905 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:01:12,350 INFO MainThread:3201905 [wandb_setup.py:_flush():81] Configure stats pid to 3201905 +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_150112-al2s20a2/logs/debug.log +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_150112-al2s20a2/logs/debug-internal.log +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:01:12,351 INFO MainThread:3201905 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:01:12,352 INFO MainThread:3201905 [wandb_init.py:init():854] starting backend +2025-09-16 15:01:12,572 INFO MainThread:3201905 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:01:12,578 INFO MainThread:3201905 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:01:12,581 INFO MainThread:3201905 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:01:12,592 INFO MainThread:3201905 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:01:13,400 INFO MainThread:3201905 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:01:13,586 INFO MainThread:3201905 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:01:13,586 INFO MainThread:3201905 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:01:13,586 INFO MainThread:3201905 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:01:13,586 INFO MainThread:3201905 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:01:13,589 INFO MainThread:3201905 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:02:28,890 INFO MainThread:3201905 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:03:41,227 INFO wandb-AsyncioManager-main:3201905 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:03:41,228 INFO wandb-AsyncioManager-main:3201905 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_150112-al2s20a2/run-al2s20a2.wandb b/wandb/run-20250916_150112-al2s20a2/run-al2s20a2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..5c7b06f3673fbaea5f71d08fe14dace2f06507c1 Binary files /dev/null and b/wandb/run-20250916_150112-al2s20a2/run-al2s20a2.wandb differ diff --git a/wandb/run-20250916_150408-f7vg4j6t/files/output.log b/wandb/run-20250916_150408-f7vg4j6t/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b16913572c86e627d81cae5e838a07277efc9146 --- /dev/null +++ b/wandb/run-20250916_150408-f7vg4j6t/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 1%| | 7/826 [03:11<6:13:53, 0.04it/s, v_num=4j6t] diff --git a/wandb/run-20250916_150408-f7vg4j6t/files/requirements.txt b/wandb/run-20250916_150408-f7vg4j6t/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_150408-f7vg4j6t/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_150408-f7vg4j6t/files/wandb-metadata.json b/wandb/run-20250916_150408-f7vg4j6t/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..24772ca87f0e3350371337e20f843c94f1565205 --- /dev/null +++ b/wandb/run-20250916_150408-f7vg4j6t/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:04:08.697942Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577045983232" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "go5meg8u07tqyqekwnif93ah1skz6zbv" +} \ No newline at end of file diff --git a/wandb/run-20250916_150408-f7vg4j6t/logs/debug-core.log b/wandb/run-20250916_150408-f7vg4j6t/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..c559134798285ded157b3d5b922bcbb1aa4a799d --- /dev/null +++ b/wandb/run-20250916_150408-f7vg4j6t/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:04:08.73293458+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpm6cpl_p5/port-3208682.txt","pid":3208682,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:04:08.733166107+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:04:08.73385715+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3208682} +{"time":"2025-09-16T15:04:08.733830304+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3208682-3209608-2716096802/socket","Net":"unix"}} +{"time":"2025-09-16T15:04:08.922427105+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:04:08.929921208+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"f7vg4j6t","id":"1(@)"} +{"time":"2025-09-16T15:04:09.408464963+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"f7vg4j6t","id":"1(@)"} +{"time":"2025-09-16T15:09:23.983708661+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_150408-f7vg4j6t/logs/debug-internal.log b/wandb/run-20250916_150408-f7vg4j6t/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..653e27a7ae76900d6c38e28f880db617eb141196 --- /dev/null +++ b/wandb/run-20250916_150408-f7vg4j6t/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:04:08.930093792+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:04:08.957112095+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:04:09.408411121+08:00","level":"INFO","msg":"stream: created new stream","id":"f7vg4j6t"} +{"time":"2025-09-16T15:04:09.408458244+08:00","level":"INFO","msg":"stream: started","id":"f7vg4j6t"} +{"time":"2025-09-16T15:04:09.408498984+08:00","level":"INFO","msg":"writer: started","stream_id":"f7vg4j6t"} +{"time":"2025-09-16T15:04:09.408531984+08:00","level":"INFO","msg":"handler: started","stream_id":"f7vg4j6t"} +{"time":"2025-09-16T15:04:09.408562565+08:00","level":"INFO","msg":"sender: started","stream_id":"f7vg4j6t"} diff --git a/wandb/run-20250916_150408-f7vg4j6t/logs/debug.log b/wandb/run-20250916_150408-f7vg4j6t/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e1d8aa3c32e66ce7ebca8c90c811bc6c6cb5e55e --- /dev/null +++ b/wandb/run-20250916_150408-f7vg4j6t/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 15:04:08,701 INFO MainThread:3208682 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_setup.py:_flush():81] Configure stats pid to 3208682 +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_150408-f7vg4j6t/logs/debug.log +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_150408-f7vg4j6t/logs/debug-internal.log +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:04:08,702 INFO MainThread:3208682 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:04:08,703 INFO MainThread:3208682 [wandb_init.py:init():854] starting backend +2025-09-16 15:04:08,922 INFO MainThread:3208682 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:04:08,927 INFO MainThread:3208682 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:04:08,930 INFO MainThread:3208682 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:04:08,940 INFO MainThread:3208682 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:04:09,741 INFO MainThread:3208682 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:04:09,953 INFO MainThread:3208682 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:04:09,953 INFO MainThread:3208682 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:04:09,953 INFO MainThread:3208682 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:04:09,953 INFO MainThread:3208682 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:04:09,955 INFO MainThread:3208682 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:05:23,815 INFO MainThread:3208682 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_150408-f7vg4j6t/run-f7vg4j6t.wandb b/wandb/run-20250916_150408-f7vg4j6t/run-f7vg4j6t.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_150945-1layq360/files/config.yaml b/wandb/run-20250916_150945-1layq360/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7a1e2bd46afc6d9bc1a2220285cab4bc5ba0d42f --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + kwkbbxjwpyubmcaatghg87ngxbizgvvg: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577046601728" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:09:45.874996Z" + writerId: kwkbbxjwpyubmcaatghg87ngxbizgvvg + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_150945-1layq360/files/output.log b/wandb/run-20250916_150945-1layq360/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_150945-1layq360/files/requirements.txt b/wandb/run-20250916_150945-1layq360/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_150945-1layq360/files/wandb-metadata.json b/wandb/run-20250916_150945-1layq360/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..864c6b7f240ac2f791d5eabf15cc7060900e8688 --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:09:45.874996Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577046601728" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "kwkbbxjwpyubmcaatghg87ngxbizgvvg" +} \ No newline at end of file diff --git a/wandb/run-20250916_150945-1layq360/files/wandb-summary.json b/wandb/run-20250916_150945-1layq360/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..6f007b0dea49eed57028e1b6f4e884b94240d34e --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":22},"_runtime":22} \ No newline at end of file diff --git a/wandb/run-20250916_150945-1layq360/logs/debug-core.log b/wandb/run-20250916_150945-1layq360/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..1b017ac4e9d8a07ac876633719f96e12526d3024 --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-16T15:09:45.912462086+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp2e5wnbvy/port-3223392.txt","pid":3223392,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:09:45.912660567+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:09:45.913499528+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3223392} +{"time":"2025-09-16T15:09:45.913462946+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3223392-3224360-1833525413/socket","Net":"unix"}} +{"time":"2025-09-16T15:09:46.100024377+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:09:46.108870243+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"1layq360","id":"1(@)"} +{"time":"2025-09-16T15:09:46.568899757+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"1layq360","id":"1(@)"} +{"time":"2025-09-16T15:10:09.511529148+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:10:09.511626784+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:10:09.511657026+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:10:09.511701369+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:10:09.512003693+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3223392-3224360-1833525413/socket","Net":"unix"}} +{"time":"2025-09-16T15:10:10.090739247+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_150945-1layq360/logs/debug-internal.log b/wandb/run-20250916_150945-1layq360/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..c8095b5f516a04152b2eb1f3969c231ddb40e0ab --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:09:46.109033263+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:09:46.134729098+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:09:46.568741067+08:00","level":"INFO","msg":"stream: created new stream","id":"1layq360"} +{"time":"2025-09-16T15:09:46.568881216+08:00","level":"INFO","msg":"stream: started","id":"1layq360"} +{"time":"2025-09-16T15:09:46.568903667+08:00","level":"INFO","msg":"writer: started","stream_id":"1layq360"} +{"time":"2025-09-16T15:09:46.568917105+08:00","level":"INFO","msg":"handler: started","stream_id":"1layq360"} +{"time":"2025-09-16T15:09:46.568944927+08:00","level":"INFO","msg":"sender: started","stream_id":"1layq360"} +{"time":"2025-09-16T15:10:09.51163723+08:00","level":"INFO","msg":"stream: closing","id":"1layq360"} diff --git a/wandb/run-20250916_150945-1layq360/logs/debug.log b/wandb/run-20250916_150945-1layq360/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c3dfe99fd14053822e3cb04f7fd2e419747e344c --- /dev/null +++ b/wandb/run-20250916_150945-1layq360/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 15:09:45,879 INFO MainThread:3223392 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:09:45,879 INFO MainThread:3223392 [wandb_setup.py:_flush():81] Configure stats pid to 3223392 +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_150945-1layq360/logs/debug.log +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_150945-1layq360/logs/debug-internal.log +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:09:45,880 INFO MainThread:3223392 [wandb_init.py:init():854] starting backend +2025-09-16 15:09:46,100 INFO MainThread:3223392 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:09:46,105 INFO MainThread:3223392 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:09:46,108 INFO MainThread:3223392 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:09:46,119 INFO MainThread:3223392 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:09:46,944 INFO MainThread:3223392 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:09:47,114 INFO MainThread:3223392 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:09:47,115 INFO MainThread:3223392 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:09:47,116 INFO MainThread:3223392 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:09:47,116 INFO MainThread:3223392 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:09:47,120 INFO MainThread:3223392 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:10:09,511 INFO wandb-AsyncioManager-main:3223392 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:10:09,512 INFO wandb-AsyncioManager-main:3223392 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_150945-1layq360/run-1layq360.wandb b/wandb/run-20250916_150945-1layq360/run-1layq360.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_151038-i3x3xsnh/files/output.log b/wandb/run-20250916_151038-i3x3xsnh/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..809066e3422c539c9a40c019551ae17e2ccc8dab --- /dev/null +++ b/wandb/run-20250916_151038-i3x3xsnh/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 1%| | 7/826 [03:15<6:21:09, 0.04it/s, v_num=xsnh] diff --git a/wandb/run-20250916_151038-i3x3xsnh/files/requirements.txt b/wandb/run-20250916_151038-i3x3xsnh/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_151038-i3x3xsnh/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_151038-i3x3xsnh/files/wandb-metadata.json b/wandb/run-20250916_151038-i3x3xsnh/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ab3ca8a594d43b86879ed74424b96a5551b6cb28 --- /dev/null +++ b/wandb/run-20250916_151038-i3x3xsnh/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:10:38.790086Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577039646720" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "e1kalcfmvljrraianju2sepgalzowx3o" +} \ No newline at end of file diff --git a/wandb/run-20250916_151038-i3x3xsnh/logs/debug-core.log b/wandb/run-20250916_151038-i3x3xsnh/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..600d05446c98be5801b66cceb447e4edcaa8e392 --- /dev/null +++ b/wandb/run-20250916_151038-i3x3xsnh/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:10:38.822789822+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp5vxquka7/port-3226353.txt","pid":3226353,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:10:38.822987017+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:10:38.823582653+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3226353} +{"time":"2025-09-16T15:10:38.823553311+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3226353-3227770-3251877260/socket","Net":"unix"}} +{"time":"2025-09-16T15:10:39.011318512+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:10:39.021027375+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"i3x3xsnh","id":"1(@)"} +{"time":"2025-09-16T15:10:39.500437054+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"i3x3xsnh","id":"1(@)"} +{"time":"2025-09-16T15:19:20.800761697+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_151038-i3x3xsnh/logs/debug-internal.log b/wandb/run-20250916_151038-i3x3xsnh/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..759d0c4333e8a9809800d205d0af2a697445fdc2 --- /dev/null +++ b/wandb/run-20250916_151038-i3x3xsnh/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:10:39.021304535+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:10:39.061965767+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:10:39.500314237+08:00","level":"INFO","msg":"stream: created new stream","id":"i3x3xsnh"} +{"time":"2025-09-16T15:10:39.500422076+08:00","level":"INFO","msg":"stream: started","id":"i3x3xsnh"} +{"time":"2025-09-16T15:10:39.500461931+08:00","level":"INFO","msg":"writer: started","stream_id":"i3x3xsnh"} +{"time":"2025-09-16T15:10:39.500475834+08:00","level":"INFO","msg":"handler: started","stream_id":"i3x3xsnh"} +{"time":"2025-09-16T15:10:39.500514808+08:00","level":"INFO","msg":"sender: started","stream_id":"i3x3xsnh"} diff --git a/wandb/run-20250916_151038-i3x3xsnh/logs/debug.log b/wandb/run-20250916_151038-i3x3xsnh/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c0166bda3dd8c8c68f500b7012e53a626151e0cf --- /dev/null +++ b/wandb/run-20250916_151038-i3x3xsnh/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 15:10:38,793 INFO MainThread:3226353 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:10:38,793 INFO MainThread:3226353 [wandb_setup.py:_flush():81] Configure stats pid to 3226353 +2025-09-16 15:10:38,793 INFO MainThread:3226353 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:10:38,793 INFO MainThread:3226353 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:10:38,793 INFO MainThread:3226353 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:10:38,794 INFO MainThread:3226353 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_151038-i3x3xsnh/logs/debug.log +2025-09-16 15:10:38,794 INFO MainThread:3226353 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_151038-i3x3xsnh/logs/debug-internal.log +2025-09-16 15:10:38,794 INFO MainThread:3226353 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:10:38,794 INFO MainThread:3226353 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:10:38,794 INFO MainThread:3226353 [wandb_init.py:init():854] starting backend +2025-09-16 15:10:39,011 INFO MainThread:3226353 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:10:39,016 INFO MainThread:3226353 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:10:39,019 INFO MainThread:3226353 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:10:39,029 INFO MainThread:3226353 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:10:39,829 INFO MainThread:3226353 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:10:39,992 INFO MainThread:3226353 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:10:39,993 INFO MainThread:3226353 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:10:39,993 INFO MainThread:3226353 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:10:39,993 INFO MainThread:3226353 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:10:39,995 INFO MainThread:3226353 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:12:05,433 INFO MainThread:3226353 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_151038-i3x3xsnh/run-i3x3xsnh.wandb b/wandb/run-20250916_151038-i3x3xsnh/run-i3x3xsnh.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_151637-dtxtymcj/files/config.yaml b/wandb/run-20250916_151637-dtxtymcj/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..54d5a336f2ed3a916026b67dcb613f6bb00bfb7a --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + esza8kth95urem8n1axco3sn9lazxqvn: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577048047616" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:16:37.312917Z" + writerId: esza8kth95urem8n1axco3sn9lazxqvn + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_151637-dtxtymcj/files/output.log b/wandb/run-20250916_151637-dtxtymcj/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..cc79235934e936308f9bf6aba1c3551af3f96747 --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/files/output.log @@ -0,0 +1,107 @@ +The length of the dataset is: 13278 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Traceback (most recent call last): + File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run + self.strategy.setup(self) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup + self.init_deepspeed() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed + self._initialize_deepspeed_train(self.model) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 496, in _initialize_deepspeed_train + model, deepspeed_optimizer = self._setup_model_and_optimizer(model, optimizer, scheduler) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 432, in _setup_model_and_optimizer + deepspeed_engine, deepspeed_optimizer, _, _ = deepspeed.initialize( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/__init__.py", line 193, in initialize + engine = DeepSpeedEngine(args=args, + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 290, in __init__ + self._configure_distributed_model(model) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 1303, in _configure_distributed_model + self.module.to(self.device) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/fabric/utilities/device_dtype_mixin.py", line 55, in to + return super().to(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1369, in to + return self._apply(convert) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 928, in _apply + module._apply(fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 928, in _apply + module._apply(fn) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 928, in _apply + module._apply(fn) + [Previous line repeated 6 more times] + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 955, in _apply + param_applied = fn(param) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1355, in convert + return t.to( +torch.AcceleratorError: CUDA error: out of memory +Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. + +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 987, in _run +[rank0]: self.strategy.setup(self) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 358, in setup +[rank0]: self.init_deepspeed() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 460, in init_deepspeed +[rank0]: self._initialize_deepspeed_train(self.model) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 496, in _initialize_deepspeed_train +[rank0]: model, deepspeed_optimizer = self._setup_model_and_optimizer(model, optimizer, scheduler) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/deepspeed.py", line 432, in _setup_model_and_optimizer +[rank0]: deepspeed_engine, deepspeed_optimizer, _, _ = deepspeed.initialize( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/__init__.py", line 193, in initialize +[rank0]: engine = DeepSpeedEngine(args=args, +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 290, in __init__ +[rank0]: self._configure_distributed_model(model) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 1303, in _configure_distributed_model +[rank0]: self.module.to(self.device) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/fabric/utilities/device_dtype_mixin.py", line 55, in to +[rank0]: return super().to(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1369, in to +[rank0]: return self._apply(convert) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 928, in _apply +[rank0]: module._apply(fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 928, in _apply +[rank0]: module._apply(fn) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 928, in _apply +[rank0]: module._apply(fn) +[rank0]: [Previous line repeated 6 more times] +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 955, in _apply +[rank0]: param_applied = fn(param) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1355, in convert +[rank0]: return t.to( +[rank0]: torch.AcceleratorError: CUDA error: out of memory +[rank0]: Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions. diff --git a/wandb/run-20250916_151637-dtxtymcj/files/requirements.txt b/wandb/run-20250916_151637-dtxtymcj/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_151637-dtxtymcj/files/wandb-metadata.json b/wandb/run-20250916_151637-dtxtymcj/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..1440e81a1088bd4fe89adda93bc51e8bd21f6c06 --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:16:37.312917Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577048047616" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "esza8kth95urem8n1axco3sn9lazxqvn" +} \ No newline at end of file diff --git a/wandb/run-20250916_151637-dtxtymcj/files/wandb-summary.json b/wandb/run-20250916_151637-dtxtymcj/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..7e8f5e934a7f744fd58fa7af47ff8a9e71aad891 --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":59},"_runtime":59} \ No newline at end of file diff --git a/wandb/run-20250916_151637-dtxtymcj/logs/debug-core.log b/wandb/run-20250916_151637-dtxtymcj/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..c413cbfedb1af44e86b3f75f93bcd3902593be10 --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:16:37.347803826+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp13n4a00b/port-3242408.txt","pid":3242408,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:16:37.347992615+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:16:37.348516085+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3242408} +{"time":"2025-09-16T15:16:37.348530883+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3242408-3244154-1101977156/socket","Net":"unix"}} +{"time":"2025-09-16T15:16:37.537435825+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:16:37.548502826+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"dtxtymcj","id":"1(@)"} +{"time":"2025-09-16T15:16:38.023991483+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"dtxtymcj","id":"1(@)"} +{"time":"2025-09-16T15:17:37.686752472+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:17:37.686837381+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:17:37.688234334+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:17:37.688417164+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:17:37.688565266+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3242408-3244154-1101977156/socket","Net":"unix"}} +{"time":"2025-09-16T15:17:39.258254696+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:17:39.258350378+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:17:39.258384784+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_151637-dtxtymcj/logs/debug-internal.log b/wandb/run-20250916_151637-dtxtymcj/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2096299e6546a7053d981eff27ec96fe3f73cfab --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:16:37.548792158+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:16:37.565086354+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:16:38.023823994+08:00","level":"INFO","msg":"stream: created new stream","id":"dtxtymcj"} +{"time":"2025-09-16T15:16:38.023967971+08:00","level":"INFO","msg":"stream: started","id":"dtxtymcj"} +{"time":"2025-09-16T15:16:38.024236947+08:00","level":"INFO","msg":"writer: started","stream_id":"dtxtymcj"} +{"time":"2025-09-16T15:16:38.02428868+08:00","level":"INFO","msg":"sender: started","stream_id":"dtxtymcj"} +{"time":"2025-09-16T15:16:38.02424956+08:00","level":"INFO","msg":"handler: started","stream_id":"dtxtymcj"} +{"time":"2025-09-16T15:17:37.688539643+08:00","level":"INFO","msg":"stream: closing","id":"dtxtymcj"} +{"time":"2025-09-16T15:17:38.990117895+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:17:39.256808932+08:00","level":"INFO","msg":"handler: closed","stream_id":"dtxtymcj"} +{"time":"2025-09-16T15:17:39.257358368+08:00","level":"INFO","msg":"sender: closed","stream_id":"dtxtymcj"} +{"time":"2025-09-16T15:17:39.257402193+08:00","level":"INFO","msg":"stream: closed","id":"dtxtymcj"} diff --git a/wandb/run-20250916_151637-dtxtymcj/logs/debug.log b/wandb/run-20250916_151637-dtxtymcj/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..29c2c46c5f16ca81df182926c5fa194963360d9c --- /dev/null +++ b/wandb/run-20250916_151637-dtxtymcj/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 15:16:37,316 INFO MainThread:3242408 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_setup.py:_flush():81] Configure stats pid to 3242408 +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_151637-dtxtymcj/logs/debug.log +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_151637-dtxtymcj/logs/debug-internal.log +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:16:37,317 INFO MainThread:3242408 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:16:37,318 INFO MainThread:3242408 [wandb_init.py:init():854] starting backend +2025-09-16 15:16:37,537 INFO MainThread:3242408 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:16:37,542 INFO MainThread:3242408 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:16:37,545 INFO MainThread:3242408 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:16:37,557 INFO MainThread:3242408 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:16:38,294 INFO MainThread:3242408 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:16:38,476 INFO MainThread:3242408 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:16:38,476 INFO MainThread:3242408 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:16:38,477 INFO MainThread:3242408 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:16:38,477 INFO MainThread:3242408 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:16:38,480 INFO MainThread:3242408 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:17:37,688 INFO wandb-AsyncioManager-main:3242408 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:17:37,688 INFO wandb-AsyncioManager-main:3242408 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_151637-dtxtymcj/run-dtxtymcj.wandb b/wandb/run-20250916_151637-dtxtymcj/run-dtxtymcj.wandb new file mode 100644 index 0000000000000000000000000000000000000000..65dd9b7555591fe280d0722b5f089cd0fcbe9961 Binary files /dev/null and b/wandb/run-20250916_151637-dtxtymcj/run-dtxtymcj.wandb differ diff --git a/wandb/run-20250916_151802-ll3gdm3e/files/config.yaml b/wandb/run-20250916_151802-ll3gdm3e/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a2b713b145e96d923912e8c0c4ed4e486696de10 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + tpfacrb6b645mt8kpt8uf12ypa3wlxej: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=8 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577049649152" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:18:02.038909Z" + writerId: tpfacrb6b645mt8kpt8uf12ypa3wlxej + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_151802-ll3gdm3e/files/output.log b/wandb/run-20250916_151802-ll3gdm3e/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7d4bb1acf0e19906951e0e06220e14ba4b604019 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/files/output.log @@ -0,0 +1,8 @@ +The length of the dataset is: 27632 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_151802-ll3gdm3e/files/requirements.txt b/wandb/run-20250916_151802-ll3gdm3e/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_151802-ll3gdm3e/files/wandb-metadata.json b/wandb/run-20250916_151802-ll3gdm3e/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..23a3dab200267079be0a47463a12b323731e4b6a --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:18:02.038909Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577049649152" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "tpfacrb6b645mt8kpt8uf12ypa3wlxej" +} \ No newline at end of file diff --git a/wandb/run-20250916_151802-ll3gdm3e/files/wandb-summary.json b/wandb/run-20250916_151802-ll3gdm3e/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..b3911042a16074700d51c6136a44fbc6ab39d905 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":50,"_wandb":{"runtime":50}} \ No newline at end of file diff --git a/wandb/run-20250916_151802-ll3gdm3e/logs/debug-core.log b/wandb/run-20250916_151802-ll3gdm3e/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..1c73c4db37dd29577cfdc54eb4b5b95f7c7f42c1 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:18:02.073601635+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp6euri14j/port-3247841.txt","pid":3247841,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:18:02.073895739+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:18:02.074664478+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3247841} +{"time":"2025-09-16T15:18:02.07467018+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3247841-3249168-1150434512/socket","Net":"unix"}} +{"time":"2025-09-16T15:18:02.263290392+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:18:02.272400732+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ll3gdm3e","id":"1(@)"} +{"time":"2025-09-16T15:18:02.725662786+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ll3gdm3e","id":"1(@)"} +{"time":"2025-09-16T15:18:53.077814304+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:18:53.077956421+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:18:53.07799119+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:18:53.078042537+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:18:53.078373771+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3247841-3249168-1150434512/socket","Net":"unix"}} +{"time":"2025-09-16T15:18:54.590813053+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:18:54.59087372+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:18:54.590919933+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_151802-ll3gdm3e/logs/debug-internal.log b/wandb/run-20250916_151802-ll3gdm3e/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0c20d3d48dbab3d80b28fc936d07915f9e8a3543 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:18:02.272723445+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:18:02.288297417+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:18:02.725537261+08:00","level":"INFO","msg":"stream: created new stream","id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:02.725648158+08:00","level":"INFO","msg":"stream: started","id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:02.725697126+08:00","level":"INFO","msg":"handler: started","stream_id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:02.725676653+08:00","level":"INFO","msg":"writer: started","stream_id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:02.72575368+08:00","level":"INFO","msg":"sender: started","stream_id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:53.077968217+08:00","level":"INFO","msg":"stream: closing","id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:54.328236366+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:18:54.589145436+08:00","level":"INFO","msg":"handler: closed","stream_id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:54.58967034+08:00","level":"INFO","msg":"sender: closed","stream_id":"ll3gdm3e"} +{"time":"2025-09-16T15:18:54.589726244+08:00","level":"INFO","msg":"stream: closed","id":"ll3gdm3e"} diff --git a/wandb/run-20250916_151802-ll3gdm3e/logs/debug.log b/wandb/run-20250916_151802-ll3gdm3e/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..810568fa6f9c5f44aefe32a4ba8695e76f8a3a89 --- /dev/null +++ b/wandb/run-20250916_151802-ll3gdm3e/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 15:18:02,042 INFO MainThread:3247841 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_setup.py:_flush():81] Configure stats pid to 3247841 +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_151802-ll3gdm3e/logs/debug.log +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_151802-ll3gdm3e/logs/debug-internal.log +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:18:02,043 INFO MainThread:3247841 [wandb_init.py:init():854] starting backend +2025-09-16 15:18:02,263 INFO MainThread:3247841 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:18:02,266 INFO MainThread:3247841 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:18:02,268 INFO MainThread:3247841 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:18:02,274 INFO MainThread:3247841 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:18:02,996 INFO MainThread:3247841 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:18:03,169 INFO MainThread:3247841 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:18:03,169 INFO MainThread:3247841 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:18:03,169 INFO MainThread:3247841 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:18:03,169 INFO MainThread:3247841 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:18:03,172 INFO MainThread:3247841 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:18:53,078 INFO wandb-AsyncioManager-main:3247841 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:18:53,078 INFO wandb-AsyncioManager-main:3247841 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_151802-ll3gdm3e/run-ll3gdm3e.wandb b/wandb/run-20250916_151802-ll3gdm3e/run-ll3gdm3e.wandb new file mode 100644 index 0000000000000000000000000000000000000000..29b7fa2417f872162c49e6aa84719cb5025907ab Binary files /dev/null and b/wandb/run-20250916_151802-ll3gdm3e/run-ll3gdm3e.wandb differ diff --git a/wandb/run-20250916_152018-7c7q5cv8/files/output.log b/wandb/run-20250916_152018-7c7q5cv8/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..eb03ed596f8d9f460b72039a55c2735debc6bc38 --- /dev/null +++ b/wandb/run-20250916_152018-7c7q5cv8/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 1%| | 7/826 [03:16<6:22:35, 0.04it/s, v_num=5cv8] diff --git a/wandb/run-20250916_152018-7c7q5cv8/files/requirements.txt b/wandb/run-20250916_152018-7c7q5cv8/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_152018-7c7q5cv8/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_152018-7c7q5cv8/files/wandb-metadata.json b/wandb/run-20250916_152018-7c7q5cv8/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..83faf3fcaeb316c09295ef61715f51e8d5a4626a --- /dev/null +++ b/wandb/run-20250916_152018-7c7q5cv8/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:20:18.484702Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=8", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577048481792" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "sc28kq1899jk0z1nuumwyjt7wotfr7zm" +} \ No newline at end of file diff --git a/wandb/run-20250916_152018-7c7q5cv8/logs/debug-core.log b/wandb/run-20250916_152018-7c7q5cv8/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..a37135f9be21e3721196df27ac54b7228dd928b6 --- /dev/null +++ b/wandb/run-20250916_152018-7c7q5cv8/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:20:18.518028766+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmphsghhgf0/port-3254965.txt","pid":3254965,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:20:18.518244256+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:20:18.518885587+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3254965} +{"time":"2025-09-16T15:20:18.518880215+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3254965-3255878-2639567454/socket","Net":"unix"}} +{"time":"2025-09-16T15:20:18.707219962+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:20:18.713356246+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"7c7q5cv8","id":"1(@)"} +{"time":"2025-09-16T15:20:19.182891684+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"7c7q5cv8","id":"1(@)"} +{"time":"2025-09-16T15:25:39.485277575+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_152018-7c7q5cv8/logs/debug-internal.log b/wandb/run-20250916_152018-7c7q5cv8/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..6f407fcfd6d3f1c142af412368f233c4d6e64102 --- /dev/null +++ b/wandb/run-20250916_152018-7c7q5cv8/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:20:18.713541663+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:20:18.738736899+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:20:19.182804046+08:00","level":"INFO","msg":"stream: created new stream","id":"7c7q5cv8"} +{"time":"2025-09-16T15:20:19.182880568+08:00","level":"INFO","msg":"stream: started","id":"7c7q5cv8"} +{"time":"2025-09-16T15:20:19.182909387+08:00","level":"INFO","msg":"writer: started","stream_id":"7c7q5cv8"} +{"time":"2025-09-16T15:20:19.182951667+08:00","level":"INFO","msg":"sender: started","stream_id":"7c7q5cv8"} +{"time":"2025-09-16T15:20:19.182915297+08:00","level":"INFO","msg":"handler: started","stream_id":"7c7q5cv8"} diff --git a/wandb/run-20250916_152018-7c7q5cv8/logs/debug.log b/wandb/run-20250916_152018-7c7q5cv8/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c86ac8987d6b4b3292029885000aa8d20b3aa9d2 --- /dev/null +++ b/wandb/run-20250916_152018-7c7q5cv8/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 15:20:18,487 INFO MainThread:3254965 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_setup.py:_flush():81] Configure stats pid to 3254965 +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_152018-7c7q5cv8/logs/debug.log +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_152018-7c7q5cv8/logs/debug-internal.log +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:20:18,488 INFO MainThread:3254965 [wandb_init.py:init():854] starting backend +2025-09-16 15:20:18,707 INFO MainThread:3254965 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:20:18,710 INFO MainThread:3254965 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:20:18,713 INFO MainThread:3254965 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:20:18,722 INFO MainThread:3254965 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:20:19,537 INFO MainThread:3254965 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:20:19,711 INFO MainThread:3254965 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:20:19,711 INFO MainThread:3254965 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:20:19,711 INFO MainThread:3254965 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:20:19,712 INFO MainThread:3254965 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:20:19,715 INFO MainThread:3254965 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:21:34,129 INFO MainThread:3254965 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_152018-7c7q5cv8/run-7c7q5cv8.wandb b/wandb/run-20250916_152018-7c7q5cv8/run-7c7q5cv8.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_152617-fpgatsd1/files/config.yaml b/wandb/run-20250916_152617-fpgatsd1/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5bd1922af99521893ca537a0787d83747223d186 --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 9qg9xkzhafff98lueexpthtc0c8t8mvy: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=7 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577049210880" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:26:17.661603Z" + writerId: 9qg9xkzhafff98lueexpthtc0c8t8mvy + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_152617-fpgatsd1/files/output.log b/wandb/run-20250916_152617-fpgatsd1/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e1af3e0eb930300b64443598f3b37f41b262ae52 --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/files/output.log @@ -0,0 +1,12 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 1/944 [00:39<10:27:09, 0.03it/s, v_num=tsd1] + +Detected KeyboardInterrupt, attempting graceful shutdown ... +[rank: 0] Received SIGTERM: 15 diff --git a/wandb/run-20250916_152617-fpgatsd1/files/requirements.txt b/wandb/run-20250916_152617-fpgatsd1/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_152617-fpgatsd1/files/wandb-metadata.json b/wandb/run-20250916_152617-fpgatsd1/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bb82f3655892141858970faf06f71970830a92ab --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:26:17.661603Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=7", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577049210880" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "9qg9xkzhafff98lueexpthtc0c8t8mvy" +} \ No newline at end of file diff --git a/wandb/run-20250916_152617-fpgatsd1/files/wandb-summary.json b/wandb/run-20250916_152617-fpgatsd1/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..828ae1cea6a02df9f9147f257c7e1634c71793e7 --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":140},"_runtime":140} \ No newline at end of file diff --git a/wandb/run-20250916_152617-fpgatsd1/logs/debug-core.log b/wandb/run-20250916_152617-fpgatsd1/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..3044094fef86d4876ce9e96da2e1f53d1405416c --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:26:17.694505975+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpm65apexi/port-3270734.txt","pid":3270734,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:26:17.694717412+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:26:17.695282027+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3270734} +{"time":"2025-09-16T15:26:17.695289854+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3270734-3271674-2677607316/socket","Net":"unix"}} +{"time":"2025-09-16T15:26:17.884280157+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:26:17.892155318+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"fpgatsd1","id":"1(@)"} +{"time":"2025-09-16T15:26:18.344625301+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"fpgatsd1","id":"1(@)"} +{"time":"2025-09-16T15:28:39.452846999+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:28:39.452954576+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:28:39.452941887+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:28:39.453032424+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:28:39.453172626+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3270734-3271674-2677607316/socket","Net":"unix"}} +{"time":"2025-09-16T15:28:40.966517576+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:28:40.966562125+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:28:40.966582029+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_152617-fpgatsd1/logs/debug-internal.log b/wandb/run-20250916_152617-fpgatsd1/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..1cdfafd87dbcc9109ab523b7b58822d71f0a1508 --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:26:17.892319156+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:26:17.913315895+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:26:18.34451874+08:00","level":"INFO","msg":"stream: created new stream","id":"fpgatsd1"} +{"time":"2025-09-16T15:26:18.344609972+08:00","level":"INFO","msg":"stream: started","id":"fpgatsd1"} +{"time":"2025-09-16T15:26:18.344779338+08:00","level":"INFO","msg":"sender: started","stream_id":"fpgatsd1"} +{"time":"2025-09-16T15:26:18.344765607+08:00","level":"INFO","msg":"writer: started","stream_id":"fpgatsd1"} +{"time":"2025-09-16T15:26:18.344816122+08:00","level":"INFO","msg":"handler: started","stream_id":"fpgatsd1"} +{"time":"2025-09-16T15:28:39.452947184+08:00","level":"INFO","msg":"stream: closing","id":"fpgatsd1"} +{"time":"2025-09-16T15:28:40.709874216+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:28:40.964034497+08:00","level":"INFO","msg":"handler: closed","stream_id":"fpgatsd1"} +{"time":"2025-09-16T15:28:40.964893737+08:00","level":"INFO","msg":"sender: closed","stream_id":"fpgatsd1"} +{"time":"2025-09-16T15:28:40.964952626+08:00","level":"INFO","msg":"stream: closed","id":"fpgatsd1"} diff --git a/wandb/run-20250916_152617-fpgatsd1/logs/debug.log b/wandb/run-20250916_152617-fpgatsd1/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..df3051402621c40e51d01fa5cb3c16181e018c49 --- /dev/null +++ b/wandb/run-20250916_152617-fpgatsd1/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:26:17,663 INFO MainThread:3270734 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_setup.py:_flush():81] Configure stats pid to 3270734 +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_152617-fpgatsd1/logs/debug.log +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_152617-fpgatsd1/logs/debug-internal.log +2025-09-16 15:26:17,664 INFO MainThread:3270734 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:26:17,665 INFO MainThread:3270734 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:26:17,665 INFO MainThread:3270734 [wandb_init.py:init():854] starting backend +2025-09-16 15:26:17,884 INFO MainThread:3270734 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:26:17,888 INFO MainThread:3270734 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:26:17,891 INFO MainThread:3270734 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:26:17,902 INFO MainThread:3270734 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:26:18,663 INFO MainThread:3270734 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:26:18,845 INFO MainThread:3270734 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:26:18,845 INFO MainThread:3270734 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:26:18,845 INFO MainThread:3270734 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:26:18,845 INFO MainThread:3270734 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:26:18,848 INFO MainThread:3270734 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:27:31,728 INFO MainThread:3270734 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:28:39,452 INFO wandb-AsyncioManager-main:3270734 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:28:39,453 INFO wandb-AsyncioManager-main:3270734 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_152617-fpgatsd1/run-fpgatsd1.wandb b/wandb/run-20250916_152617-fpgatsd1/run-fpgatsd1.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e4e91867bf1c5e785ca64135d75b399cc3d88c3e Binary files /dev/null and b/wandb/run-20250916_152617-fpgatsd1/run-fpgatsd1.wandb differ diff --git a/wandb/run-20250916_152908-2jyzyvjj/files/output.log b/wandb/run-20250916_152908-2jyzyvjj/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e5c216c913fb381b976b042b1e89d68ddc7c4e98 --- /dev/null +++ b/wandb/run-20250916_152908-2jyzyvjj/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3,4,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 1/944 [00:40<10:30:13, 0.02it/s, v_num=yvjj] diff --git a/wandb/run-20250916_152908-2jyzyvjj/files/requirements.txt b/wandb/run-20250916_152908-2jyzyvjj/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_152908-2jyzyvjj/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_152908-2jyzyvjj/files/wandb-metadata.json b/wandb/run-20250916_152908-2jyzyvjj/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..924f10033196b23e10906dd072e567d67d1aee98 --- /dev/null +++ b/wandb/run-20250916_152908-2jyzyvjj/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:29:08.885278Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=7", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577049575424" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "u4o7clmn6w72zuohaxk22l4dpidxvpfn" +} \ No newline at end of file diff --git a/wandb/run-20250916_152908-2jyzyvjj/logs/debug-core.log b/wandb/run-20250916_152908-2jyzyvjj/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..589d984995a9e89e017329fc316a318d9c3cc083 --- /dev/null +++ b/wandb/run-20250916_152908-2jyzyvjj/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:29:08.921802291+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp572i1tj1/port-3279715.txt","pid":3279715,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:29:08.922020635+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:29:08.922604667+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3279715} +{"time":"2025-09-16T15:29:08.922577476+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3279715-3280997-1256718109/socket","Net":"unix"}} +{"time":"2025-09-16T15:29:09.110064295+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:29:09.11780465+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"2jyzyvjj","id":"1(@)"} +{"time":"2025-09-16T15:29:09.579957229+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"2jyzyvjj","id":"1(@)"} +{"time":"2025-09-16T15:32:05.868385254+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_152908-2jyzyvjj/logs/debug-internal.log b/wandb/run-20250916_152908-2jyzyvjj/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e06b69d5043a2f4bb5ccbf6d5d8183227f034e49 --- /dev/null +++ b/wandb/run-20250916_152908-2jyzyvjj/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:29:09.117956105+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:29:09.142242861+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:29:09.579831956+08:00","level":"INFO","msg":"stream: created new stream","id":"2jyzyvjj"} +{"time":"2025-09-16T15:29:09.579942607+08:00","level":"INFO","msg":"stream: started","id":"2jyzyvjj"} +{"time":"2025-09-16T15:29:09.579957674+08:00","level":"INFO","msg":"writer: started","stream_id":"2jyzyvjj"} +{"time":"2025-09-16T15:29:09.580085796+08:00","level":"INFO","msg":"sender: started","stream_id":"2jyzyvjj"} +{"time":"2025-09-16T15:29:09.580126406+08:00","level":"INFO","msg":"handler: started","stream_id":"2jyzyvjj"} diff --git a/wandb/run-20250916_152908-2jyzyvjj/logs/debug.log b/wandb/run-20250916_152908-2jyzyvjj/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e18501bb450e0c9952924eaf593dea7a848afcf0 --- /dev/null +++ b/wandb/run-20250916_152908-2jyzyvjj/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 15:29:08,889 INFO MainThread:3279715 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:29:08,889 INFO MainThread:3279715 [wandb_setup.py:_flush():81] Configure stats pid to 3279715 +2025-09-16 15:29:08,889 INFO MainThread:3279715 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:29:08,889 INFO MainThread:3279715 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:29:08,889 INFO MainThread:3279715 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:29:08,889 INFO MainThread:3279715 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_152908-2jyzyvjj/logs/debug.log +2025-09-16 15:29:08,890 INFO MainThread:3279715 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_152908-2jyzyvjj/logs/debug-internal.log +2025-09-16 15:29:08,890 INFO MainThread:3279715 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:29:08,890 INFO MainThread:3279715 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:29:08,890 INFO MainThread:3279715 [wandb_init.py:init():854] starting backend +2025-09-16 15:29:09,110 INFO MainThread:3279715 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:29:09,114 INFO MainThread:3279715 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:29:09,117 INFO MainThread:3279715 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:29:09,127 INFO MainThread:3279715 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:29:09,878 INFO MainThread:3279715 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:29:10,044 INFO MainThread:3279715 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:29:10,044 INFO MainThread:3279715 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:29:10,044 INFO MainThread:3279715 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:29:10,044 INFO MainThread:3279715 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:29:10,047 INFO MainThread:3279715 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:30:22,722 INFO MainThread:3279715 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_152908-2jyzyvjj/run-2jyzyvjj.wandb b/wandb/run-20250916_152908-2jyzyvjj/run-2jyzyvjj.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_153255-oy54tkix/files/output.log b/wandb/run-20250916_153255-oy54tkix/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..243c83e2e3d951f421b699b1b09fd0aa79ca4f4c --- /dev/null +++ b/wandb/run-20250916_153255-oy54tkix/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:10:17, 0.04it/s, v_num=tkix] diff --git a/wandb/run-20250916_153255-oy54tkix/files/requirements.txt b/wandb/run-20250916_153255-oy54tkix/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_153255-oy54tkix/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_153255-oy54tkix/files/wandb-metadata.json b/wandb/run-20250916_153255-oy54tkix/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e7f6641180d8676c26c08ec70a91a58e27ce2d64 --- /dev/null +++ b/wandb/run-20250916_153255-oy54tkix/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:32:55.415200Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577050124288" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "dikmvuk8hmorfai1jlw120zqey5gjgt2" +} \ No newline at end of file diff --git a/wandb/run-20250916_153255-oy54tkix/logs/debug-core.log b/wandb/run-20250916_153255-oy54tkix/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..a85b876f500a223a3a8ef8f879117dc0a685a036 --- /dev/null +++ b/wandb/run-20250916_153255-oy54tkix/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:32:55.450678891+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpfqtcl0zj/port-3291148.txt","pid":3291148,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:32:55.4509551+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:32:55.452059441+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3291148} +{"time":"2025-09-16T15:32:55.452110629+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3291148-3291817-3071964408/socket","Net":"unix"}} +{"time":"2025-09-16T15:32:55.639993811+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:32:55.64943939+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"oy54tkix","id":"1(@)"} +{"time":"2025-09-16T15:32:56.119648601+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"oy54tkix","id":"1(@)"} +{"time":"2025-09-16T15:35:42.001250017+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_153255-oy54tkix/logs/debug-internal.log b/wandb/run-20250916_153255-oy54tkix/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a39aaeedff39fcb33049a87f61412f69ba0420e8 --- /dev/null +++ b/wandb/run-20250916_153255-oy54tkix/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:32:55.649611409+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:32:55.666366468+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:32:56.119529622+08:00","level":"INFO","msg":"stream: created new stream","id":"oy54tkix"} +{"time":"2025-09-16T15:32:56.119631791+08:00","level":"INFO","msg":"stream: started","id":"oy54tkix"} +{"time":"2025-09-16T15:32:56.11978936+08:00","level":"INFO","msg":"writer: started","stream_id":"oy54tkix"} +{"time":"2025-09-16T15:32:56.11980598+08:00","level":"INFO","msg":"sender: started","stream_id":"oy54tkix"} +{"time":"2025-09-16T15:32:56.119881572+08:00","level":"INFO","msg":"handler: started","stream_id":"oy54tkix"} diff --git a/wandb/run-20250916_153255-oy54tkix/logs/debug.log b/wandb/run-20250916_153255-oy54tkix/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..bdb23c31d44ca334fc2ca63a88c4d91b5a0ccb33 --- /dev/null +++ b/wandb/run-20250916_153255-oy54tkix/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 15:32:55,419 INFO MainThread:3291148 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:32:55,419 INFO MainThread:3291148 [wandb_setup.py:_flush():81] Configure stats pid to 3291148 +2025-09-16 15:32:55,419 INFO MainThread:3291148 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:32:55,419 INFO MainThread:3291148 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:32:55,419 INFO MainThread:3291148 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:32:55,419 INFO MainThread:3291148 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_153255-oy54tkix/logs/debug.log +2025-09-16 15:32:55,420 INFO MainThread:3291148 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_153255-oy54tkix/logs/debug-internal.log +2025-09-16 15:32:55,420 INFO MainThread:3291148 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:32:55,420 INFO MainThread:3291148 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:32:55,420 INFO MainThread:3291148 [wandb_init.py:init():854] starting backend +2025-09-16 15:32:55,640 INFO MainThread:3291148 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:32:55,644 INFO MainThread:3291148 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:32:55,647 INFO MainThread:3291148 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:32:55,658 INFO MainThread:3291148 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:32:56,486 INFO MainThread:3291148 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:32:56,651 INFO MainThread:3291148 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:32:56,651 INFO MainThread:3291148 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:32:56,651 INFO MainThread:3291148 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:32:56,651 INFO MainThread:3291148 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:32:56,654 INFO MainThread:3291148 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:34:05,308 INFO MainThread:3291148 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_153255-oy54tkix/run-oy54tkix.wandb b/wandb/run-20250916_153255-oy54tkix/run-oy54tkix.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_153604-23gwpn82/files/config.yaml b/wandb/run-20250916_153604-23gwpn82/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a593ba51bd02c0953b4ee3ccc24e04f5124117d --- /dev/null +++ b/wandb/run-20250916_153604-23gwpn82/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + avt4olmxxbfm7scz3bwo0vcy6ah1f2pz: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577050513408" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:36:04.318545Z" + writerId: avt4olmxxbfm7scz3bwo0vcy6ah1f2pz + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_153604-23gwpn82/files/output.log b/wandb/run-20250916_153604-23gwpn82/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..99b93b918b991c164a63da2728d23be339547e1d --- /dev/null +++ b/wandb/run-20250916_153604-23gwpn82/files/output.log @@ -0,0 +1,11 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/3433 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 536, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 614, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 618, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 131, in get_logsum_prob + obs_ids = self.tokenizer(observation, return_tensors="pt", padding=True).input_ids.to(self.device) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ + raise AttributeError( +AttributeError: 'ActorModel' object has no attribute 'device' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 536, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 614, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 618, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 131, in get_logsum_prob +[rank0]: obs_ids = self.tokenizer(observation, return_tensors="pt", padding=True).input_ids.to(self.device) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'ActorModel' object has no attribute 'device' diff --git a/wandb/run-20250916_153801-trilffgr/files/requirements.txt b/wandb/run-20250916_153801-trilffgr/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_153801-trilffgr/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_153801-trilffgr/files/wandb-metadata.json b/wandb/run-20250916_153801-trilffgr/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b7cf6f9ed4fa80afe77bb4714ba0ba274b943752 --- /dev/null +++ b/wandb/run-20250916_153801-trilffgr/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:38:01.493614Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577050791936" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "2pwclf9kfhyzd4euj9kbkce5pa1ynlgs" +} \ No newline at end of file diff --git a/wandb/run-20250916_153801-trilffgr/files/wandb-summary.json b/wandb/run-20250916_153801-trilffgr/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..9f39067485dc32618da84a815b6711331233c7e9 --- /dev/null +++ b/wandb/run-20250916_153801-trilffgr/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":75},"_runtime":75} \ No newline at end of file diff --git a/wandb/run-20250916_153801-trilffgr/logs/debug-core.log b/wandb/run-20250916_153801-trilffgr/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..89a599e1b2d9e4f2df809cfbe96263f2bcf954c1 --- /dev/null +++ b/wandb/run-20250916_153801-trilffgr/logs/debug-core.log @@ -0,0 +1,16 @@ +{"time":"2025-09-16T15:38:01.529628139+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpqed2n5l8/port-3305866.txt","pid":3305866,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:38:01.529849954+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:38:01.530461706+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3305866} +{"time":"2025-09-16T15:38:01.530456941+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3305866-3306577-2529349485/socket","Net":"unix"}} +{"time":"2025-09-16T15:38:01.718042156+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:38:01.727617993+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"trilffgr","id":"1(@)"} +{"time":"2025-09-16T15:38:02.21787431+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"trilffgr","id":"1(@)"} +{"time":"2025-09-16T15:39:17.892454453+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:39:17.892865521+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:39:17.892852838+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:39:17.892940354+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:39:17.893155065+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3305866-3306577-2529349485/socket","Net":"unix"}} +{"time":"2025-09-16T15:39:17.923452878+08:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write unix /tmp/wandb-3305866-3306577-2529349485/socket->@: use of closed network connection","id":"1(@)"} +{"time":"2025-09-16T15:39:19.126810546+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:39:19.126858869+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:39:19.126894139+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_153801-trilffgr/logs/debug-internal.log b/wandb/run-20250916_153801-trilffgr/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0470400a9aef1b16aff0c81bbefde3356672a204 --- /dev/null +++ b/wandb/run-20250916_153801-trilffgr/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:38:01.727864533+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:38:01.7666806+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:38:02.217587431+08:00","level":"INFO","msg":"stream: created new stream","id":"trilffgr"} +{"time":"2025-09-16T15:38:02.217858893+08:00","level":"INFO","msg":"stream: started","id":"trilffgr"} +{"time":"2025-09-16T15:38:02.217885283+08:00","level":"INFO","msg":"writer: started","stream_id":"trilffgr"} +{"time":"2025-09-16T15:38:02.2179308+08:00","level":"INFO","msg":"sender: started","stream_id":"trilffgr"} +{"time":"2025-09-16T15:38:02.21792403+08:00","level":"INFO","msg":"handler: started","stream_id":"trilffgr"} +{"time":"2025-09-16T15:39:17.892881735+08:00","level":"INFO","msg":"stream: closing","id":"trilffgr"} +{"time":"2025-09-16T15:39:18.697339689+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:39:19.12521851+08:00","level":"INFO","msg":"handler: closed","stream_id":"trilffgr"} +{"time":"2025-09-16T15:39:19.125722734+08:00","level":"INFO","msg":"sender: closed","stream_id":"trilffgr"} +{"time":"2025-09-16T15:39:19.125791801+08:00","level":"INFO","msg":"stream: closed","id":"trilffgr"} diff --git a/wandb/run-20250916_153801-trilffgr/logs/debug.log b/wandb/run-20250916_153801-trilffgr/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..8ac2af65847462af8e6c6b1a8635f753d2035009 --- /dev/null +++ b/wandb/run-20250916_153801-trilffgr/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:38:01,497 INFO MainThread:3305866 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:38:01,497 INFO MainThread:3305866 [wandb_setup.py:_flush():81] Configure stats pid to 3305866 +2025-09-16 15:38:01,497 INFO MainThread:3305866 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:38:01,497 INFO MainThread:3305866 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:38:01,498 INFO MainThread:3305866 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:38:01,498 INFO MainThread:3305866 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_153801-trilffgr/logs/debug.log +2025-09-16 15:38:01,498 INFO MainThread:3305866 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_153801-trilffgr/logs/debug-internal.log +2025-09-16 15:38:01,498 INFO MainThread:3305866 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:38:01,498 INFO MainThread:3305866 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:38:01,498 INFO MainThread:3305866 [wandb_init.py:init():854] starting backend +2025-09-16 15:38:01,718 INFO MainThread:3305866 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:38:01,722 INFO MainThread:3305866 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:38:01,725 INFO MainThread:3305866 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:38:01,737 INFO MainThread:3305866 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:38:02,549 INFO MainThread:3305866 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:38:02,671 INFO MainThread:3305866 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:38:02,672 INFO MainThread:3305866 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:38:02,672 INFO MainThread:3305866 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:38:02,672 INFO MainThread:3305866 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:38:02,675 INFO MainThread:3305866 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:39:12,580 INFO MainThread:3305866 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:39:17,892 INFO wandb-AsyncioManager-main:3305866 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:39:17,892 INFO wandb-AsyncioManager-main:3305866 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250916_153801-trilffgr/run-trilffgr.wandb b/wandb/run-20250916_153801-trilffgr/run-trilffgr.wandb new file mode 100644 index 0000000000000000000000000000000000000000..3392bfe43a72687814d47db03f83164974b21da7 Binary files /dev/null and b/wandb/run-20250916_153801-trilffgr/run-trilffgr.wandb differ diff --git a/wandb/run-20250916_153941-zhu1k0lz/files/config.yaml b/wandb/run-20250916_153941-zhu1k0lz/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..7b99828c0e7cb5f6a41dc2c31e68c42a1982e90d --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + cq4inwvya2mglu8rxisbx0ea1al8v6g4: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577051090944" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:39:41.256624Z" + writerId: cq4inwvya2mglu8rxisbx0ea1al8v6g4 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_153941-zhu1k0lz/files/output.log b/wandb/run-20250916_153941-zhu1k0lz/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..61bd2041787d1fa23177b659ba7cc5d976151042 --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/files/output.log @@ -0,0 +1,141 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/3433 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 536, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 614, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 618, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 131, in get_logsum_prob + obs_ids = self.tokenizer(observation, return_tensors="pt", padding=True).input_ids.to(self.device) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ + raise AttributeError( +AttributeError: 'ActorModel' object has no attribute 'device' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 536, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 614, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 618, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 131, in get_logsum_prob +[rank0]: obs_ids = self.tokenizer(observation, return_tensors="pt", padding=True).input_ids.to(self.device) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'ActorModel' object has no attribute 'device' diff --git a/wandb/run-20250916_153941-zhu1k0lz/files/requirements.txt b/wandb/run-20250916_153941-zhu1k0lz/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_153941-zhu1k0lz/files/wandb-metadata.json b/wandb/run-20250916_153941-zhu1k0lz/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e9509ede96e0841bfffc81b8d0476e7b35095050 --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:39:41.256624Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577051090944" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "cq4inwvya2mglu8rxisbx0ea1al8v6g4" +} \ No newline at end of file diff --git a/wandb/run-20250916_153941-zhu1k0lz/files/wandb-summary.json b/wandb/run-20250916_153941-zhu1k0lz/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..a0ca6484febe281cf369ff1fa53c341e08f4cd8f --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":74,"_wandb":{"runtime":74}} \ No newline at end of file diff --git a/wandb/run-20250916_153941-zhu1k0lz/logs/debug-core.log b/wandb/run-20250916_153941-zhu1k0lz/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..b9e1ef1234b2ad1475d92154bd2cd762cb619498 --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:39:41.291885255+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp2cd0_0y6/port-3311291.txt","pid":3311291,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:39:41.292102084+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:39:41.293813339+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3311291-3312149-2331439917/socket","Net":"unix"}} +{"time":"2025-09-16T15:39:41.29393426+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3311291} +{"time":"2025-09-16T15:39:41.481024173+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:39:41.488142466+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"zhu1k0lz","id":"1(@)"} +{"time":"2025-09-16T15:39:42.002586104+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"zhu1k0lz","id":"1(@)"} +{"time":"2025-09-16T15:40:57.250684152+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:40:57.25074973+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:40:57.250733135+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:40:57.250825715+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:40:57.250869019+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3311291-3312149-2331439917/socket","Net":"unix"}} +{"time":"2025-09-16T15:40:58.304904006+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:40:58.304945571+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:40:58.304964424+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_153941-zhu1k0lz/logs/debug-internal.log b/wandb/run-20250916_153941-zhu1k0lz/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..652c4347c4cadf412c595bdafcf6c8c6330f9f28 --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:39:41.488333433+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:39:41.516648953+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:39:42.002413606+08:00","level":"INFO","msg":"stream: created new stream","id":"zhu1k0lz"} +{"time":"2025-09-16T15:39:42.002567276+08:00","level":"INFO","msg":"stream: started","id":"zhu1k0lz"} +{"time":"2025-09-16T15:39:42.002646835+08:00","level":"INFO","msg":"writer: started","stream_id":"zhu1k0lz"} +{"time":"2025-09-16T15:39:42.002763348+08:00","level":"INFO","msg":"handler: started","stream_id":"zhu1k0lz"} +{"time":"2025-09-16T15:39:42.002790441+08:00","level":"INFO","msg":"sender: started","stream_id":"zhu1k0lz"} +{"time":"2025-09-16T15:40:57.250737531+08:00","level":"INFO","msg":"stream: closing","id":"zhu1k0lz"} +{"time":"2025-09-16T15:40:58.022573893+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:40:58.302999298+08:00","level":"INFO","msg":"handler: closed","stream_id":"zhu1k0lz"} +{"time":"2025-09-16T15:40:58.30371367+08:00","level":"INFO","msg":"sender: closed","stream_id":"zhu1k0lz"} +{"time":"2025-09-16T15:40:58.303801317+08:00","level":"INFO","msg":"stream: closed","id":"zhu1k0lz"} diff --git a/wandb/run-20250916_153941-zhu1k0lz/logs/debug.log b/wandb/run-20250916_153941-zhu1k0lz/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..5d4256b6adeeadbb865cf46e6e26fe447be62208 --- /dev/null +++ b/wandb/run-20250916_153941-zhu1k0lz/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:39:41,260 INFO MainThread:3311291 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:39:41,260 INFO MainThread:3311291 [wandb_setup.py:_flush():81] Configure stats pid to 3311291 +2025-09-16 15:39:41,260 INFO MainThread:3311291 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_153941-zhu1k0lz/logs/debug.log +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_153941-zhu1k0lz/logs/debug-internal.log +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:39:41,261 INFO MainThread:3311291 [wandb_init.py:init():854] starting backend +2025-09-16 15:39:41,481 INFO MainThread:3311291 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:39:41,485 INFO MainThread:3311291 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:39:41,487 INFO MainThread:3311291 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:39:41,496 INFO MainThread:3311291 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:39:42,342 INFO MainThread:3311291 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:39:42,505 INFO MainThread:3311291 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:39:42,505 INFO MainThread:3311291 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:39:42,507 INFO MainThread:3311291 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:39:42,507 INFO MainThread:3311291 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:39:42,510 INFO MainThread:3311291 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:40:52,076 INFO MainThread:3311291 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:40:57,250 INFO wandb-AsyncioManager-main:3311291 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:40:57,250 INFO wandb-AsyncioManager-main:3311291 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_153941-zhu1k0lz/run-zhu1k0lz.wandb b/wandb/run-20250916_153941-zhu1k0lz/run-zhu1k0lz.wandb new file mode 100644 index 0000000000000000000000000000000000000000..5f739101d6f12228c5be6da81ea0a3d1f048b124 Binary files /dev/null and b/wandb/run-20250916_153941-zhu1k0lz/run-zhu1k0lz.wandb differ diff --git a/wandb/run-20250916_154221-t0kgpbk6/files/config.yaml b/wandb/run-20250916_154221-t0kgpbk6/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a5f945d319d668b2817ea723ae84961b59cdd502 --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + txzxa4m8d8q20nt9khmbxy9apc1vex60: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577051496448" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:42:21.408728Z" + writerId: txzxa4m8d8q20nt9khmbxy9apc1vex60 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_154221-t0kgpbk6/files/output.log b/wandb/run-20250916_154221-t0kgpbk6/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..775ca55ef2143b2c46e15aeba449246e7fc09664 --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/files/output.log @@ -0,0 +1,141 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 538, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 616, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 620, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 127, in get_logsum_prob + device = next(self.actor.model.parameters()).device + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ + raise AttributeError( +AttributeError: 'ActorModel' object has no attribute 'actor' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 538, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 616, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 620, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 127, in get_logsum_prob +[rank0]: device = next(self.actor.model.parameters()).device +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'ActorModel' object has no attribute 'actor' diff --git a/wandb/run-20250916_154221-t0kgpbk6/files/requirements.txt b/wandb/run-20250916_154221-t0kgpbk6/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_154221-t0kgpbk6/files/wandb-metadata.json b/wandb/run-20250916_154221-t0kgpbk6/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..d7c54acc961dcd9622b40b32896cf36d77a26a3c --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:42:21.408728Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577051496448" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "txzxa4m8d8q20nt9khmbxy9apc1vex60" +} \ No newline at end of file diff --git a/wandb/run-20250916_154221-t0kgpbk6/files/wandb-summary.json b/wandb/run-20250916_154221-t0kgpbk6/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..14dc81afa3650cdefefbb314e5c92f7df13c46a9 --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":74},"_runtime":74} \ No newline at end of file diff --git a/wandb/run-20250916_154221-t0kgpbk6/logs/debug-core.log b/wandb/run-20250916_154221-t0kgpbk6/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..f93c1a97b80bfd82346e0363ea98e3d9f21a958f --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:42:21.44374357+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpr88z30gs/port-3318575.txt","pid":3318575,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:42:21.444026273+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:42:21.445225825+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3318575} +{"time":"2025-09-16T15:42:21.445220988+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3318575-3319293-1030242936/socket","Net":"unix"}} +{"time":"2025-09-16T15:42:21.632221638+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:42:21.641737478+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"t0kgpbk6","id":"1(@)"} +{"time":"2025-09-16T15:42:22.152101776+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"t0kgpbk6","id":"1(@)"} +{"time":"2025-09-16T15:43:36.578727548+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:43:36.578826165+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:43:36.57880889+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:43:36.578902566+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:43:36.579013469+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3318575-3319293-1030242936/socket","Net":"unix"}} +{"time":"2025-09-16T15:43:38.123486167+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:43:38.123521094+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:43:38.123539804+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_154221-t0kgpbk6/logs/debug-internal.log b/wandb/run-20250916_154221-t0kgpbk6/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..be4f2265ce1caedcc8c2d3a39da57a69649d49ef --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:42:21.641986114+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:42:21.674314832+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:42:22.151970226+08:00","level":"INFO","msg":"stream: created new stream","id":"t0kgpbk6"} +{"time":"2025-09-16T15:42:22.152087063+08:00","level":"INFO","msg":"stream: started","id":"t0kgpbk6"} +{"time":"2025-09-16T15:42:22.152120403+08:00","level":"INFO","msg":"writer: started","stream_id":"t0kgpbk6"} +{"time":"2025-09-16T15:42:22.152171753+08:00","level":"INFO","msg":"handler: started","stream_id":"t0kgpbk6"} +{"time":"2025-09-16T15:42:22.152173453+08:00","level":"INFO","msg":"sender: started","stream_id":"t0kgpbk6"} +{"time":"2025-09-16T15:43:36.578810643+08:00","level":"INFO","msg":"stream: closing","id":"t0kgpbk6"} +{"time":"2025-09-16T15:43:37.85591888+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:43:38.121455623+08:00","level":"INFO","msg":"handler: closed","stream_id":"t0kgpbk6"} +{"time":"2025-09-16T15:43:38.122023614+08:00","level":"INFO","msg":"sender: closed","stream_id":"t0kgpbk6"} +{"time":"2025-09-16T15:43:38.12208937+08:00","level":"INFO","msg":"stream: closed","id":"t0kgpbk6"} diff --git a/wandb/run-20250916_154221-t0kgpbk6/logs/debug.log b/wandb/run-20250916_154221-t0kgpbk6/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..1a1099b7ebca6df5681d897a732dcc130bc88ddd --- /dev/null +++ b/wandb/run-20250916_154221-t0kgpbk6/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:42:21,412 INFO MainThread:3318575 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_setup.py:_flush():81] Configure stats pid to 3318575 +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_154221-t0kgpbk6/logs/debug.log +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_154221-t0kgpbk6/logs/debug-internal.log +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:42:21,413 INFO MainThread:3318575 [wandb_init.py:init():854] starting backend +2025-09-16 15:42:21,632 INFO MainThread:3318575 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:42:21,637 INFO MainThread:3318575 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:42:21,639 INFO MainThread:3318575 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:42:21,651 INFO MainThread:3318575 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:42:22,532 INFO MainThread:3318575 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:42:22,692 INFO MainThread:3318575 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:42:22,692 INFO MainThread:3318575 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:42:22,693 INFO MainThread:3318575 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:42:22,693 INFO MainThread:3318575 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:42:22,695 INFO MainThread:3318575 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:43:31,674 INFO MainThread:3318575 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:43:36,578 INFO wandb-AsyncioManager-main:3318575 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:43:36,579 INFO wandb-AsyncioManager-main:3318575 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_154221-t0kgpbk6/run-t0kgpbk6.wandb b/wandb/run-20250916_154221-t0kgpbk6/run-t0kgpbk6.wandb new file mode 100644 index 0000000000000000000000000000000000000000..31928a798f4cc94ec88b127a1492a85f7a32f439 Binary files /dev/null and b/wandb/run-20250916_154221-t0kgpbk6/run-t0kgpbk6.wandb differ diff --git a/wandb/run-20250916_154401-cygxnwrm/files/config.yaml b/wandb/run-20250916_154401-cygxnwrm/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fcaa2f44a9d1ea3159cf0c9dc333869368d1c818 --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + wt5poxanxjttpmwphbgtk0wmjqg9s4rp: + args: + - fit + - --data=StrategicDialogue + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=Strategic-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577051774976" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:44:01.016295Z" + writerId: wt5poxanxjttpmwphbgtk0wmjqg9s4rp + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_154401-cygxnwrm/files/output.log b/wandb/run-20250916_154401-cygxnwrm/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_154401-cygxnwrm/files/requirements.txt b/wandb/run-20250916_154401-cygxnwrm/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_154401-cygxnwrm/files/wandb-metadata.json b/wandb/run-20250916_154401-cygxnwrm/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6971dd5a94c714772af023751244ac94e46ed0f8 --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:44:01.016295Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577051774976" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "wt5poxanxjttpmwphbgtk0wmjqg9s4rp" +} \ No newline at end of file diff --git a/wandb/run-20250916_154401-cygxnwrm/files/wandb-summary.json b/wandb/run-20250916_154401-cygxnwrm/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..1af1ec3940714bd273a10f70a8eb86a7ec99ff3f --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":15},"_runtime":15} \ No newline at end of file diff --git a/wandb/run-20250916_154401-cygxnwrm/logs/debug-core.log b/wandb/run-20250916_154401-cygxnwrm/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..cde7b7ae5b77175d8117784dad43585bf5476668 --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-16T15:44:01.054448439+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp84hyuxnp/port-3323534.txt","pid":3323534,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:44:01.054622192+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:44:01.055172232+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3323534} +{"time":"2025-09-16T15:44:01.055180524+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3323534-3324271-1815428445/socket","Net":"unix"}} +{"time":"2025-09-16T15:44:01.242816803+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:44:01.259226696+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"cygxnwrm","id":"1(@)"} +{"time":"2025-09-16T15:44:02.050351915+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"cygxnwrm","id":"1(@)"} +{"time":"2025-09-16T15:44:18.425508339+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:44:18.42560915+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:44:18.425589133+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:44:18.425727216+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:44:18.425914988+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3323534-3324271-1815428445/socket","Net":"unix"}} +{"time":"2025-09-16T15:44:18.928845558+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_154401-cygxnwrm/logs/debug-internal.log b/wandb/run-20250916_154401-cygxnwrm/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..f3844dce9dd0126ab2e9079670952df03f39e60a --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:44:01.259450276+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:44:01.300659235+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:44:02.050256192+08:00","level":"INFO","msg":"stream: created new stream","id":"cygxnwrm"} +{"time":"2025-09-16T15:44:02.050336255+08:00","level":"INFO","msg":"stream: started","id":"cygxnwrm"} +{"time":"2025-09-16T15:44:02.05036853+08:00","level":"INFO","msg":"sender: started","stream_id":"cygxnwrm"} +{"time":"2025-09-16T15:44:02.050367195+08:00","level":"INFO","msg":"writer: started","stream_id":"cygxnwrm"} +{"time":"2025-09-16T15:44:02.050445362+08:00","level":"INFO","msg":"handler: started","stream_id":"cygxnwrm"} +{"time":"2025-09-16T15:44:18.425573358+08:00","level":"INFO","msg":"stream: closing","id":"cygxnwrm"} diff --git a/wandb/run-20250916_154401-cygxnwrm/logs/debug.log b/wandb/run-20250916_154401-cygxnwrm/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..d98784fb3bd496906652673d683ee9c890c63bc0 --- /dev/null +++ b/wandb/run-20250916_154401-cygxnwrm/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 15:44:01,020 INFO MainThread:3323534 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:44:01,020 INFO MainThread:3323534 [wandb_setup.py:_flush():81] Configure stats pid to 3323534 +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_154401-cygxnwrm/logs/debug.log +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_154401-cygxnwrm/logs/debug-internal.log +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:44:01,021 INFO MainThread:3323534 [wandb_init.py:init():854] starting backend +2025-09-16 15:44:01,243 INFO MainThread:3323534 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:44:01,248 INFO MainThread:3323534 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:44:01,256 INFO MainThread:3323534 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:44:01,268 INFO MainThread:3323534 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:44:02,499 INFO MainThread:3323534 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:44:02,622 INFO MainThread:3323534 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:44:02,622 INFO MainThread:3323534 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:44:02,622 INFO MainThread:3323534 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:44:02,622 INFO MainThread:3323534 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:44:02,625 INFO MainThread:3323534 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:44:18,425 INFO wandb-AsyncioManager-main:3323534 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:44:18,425 INFO wandb-AsyncioManager-main:3323534 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_154401-cygxnwrm/run-cygxnwrm.wandb b/wandb/run-20250916_154401-cygxnwrm/run-cygxnwrm.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_154446-2dn97uuu/files/config.yaml b/wandb/run-20250916_154446-2dn97uuu/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f5e0450776b0f835007fe4d4a05289a2349683f1 --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 1rtv0z2ksyummik7h302o9km2s7bqyjv: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577051992064" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:44:46.209360Z" + writerId: 1rtv0z2ksyummik7h302o9km2s7bqyjv + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_154446-2dn97uuu/files/output.log b/wandb/run-20250916_154446-2dn97uuu/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..2f10712a0a12a9253be6371d224c3b172acd52a4 --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/files/output.log @@ -0,0 +1,141 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 538, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 616, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 620, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 133, in get_logsum_prob + obs_ids = self.tokenizer(observation, return_tensors="pt", padding=True).input_ids.to(self.device) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ + raise AttributeError( +AttributeError: 'ActorModel' object has no attribute 'device' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 538, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 616, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 620, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 133, in get_logsum_prob +[rank0]: obs_ids = self.tokenizer(observation, return_tensors="pt", padding=True).input_ids.to(self.device) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'ActorModel' object has no attribute 'device' diff --git a/wandb/run-20250916_154446-2dn97uuu/files/requirements.txt b/wandb/run-20250916_154446-2dn97uuu/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_154446-2dn97uuu/files/wandb-metadata.json b/wandb/run-20250916_154446-2dn97uuu/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b7d414fa11f675bbca910d71fa4e442f699c1eec --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:44:46.209360Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577051992064" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "1rtv0z2ksyummik7h302o9km2s7bqyjv" +} \ No newline at end of file diff --git a/wandb/run-20250916_154446-2dn97uuu/files/wandb-summary.json b/wandb/run-20250916_154446-2dn97uuu/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..65efeb016520ebca12e5019029f5a6f6602702bd --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":73},"_runtime":73} \ No newline at end of file diff --git a/wandb/run-20250916_154446-2dn97uuu/logs/debug-core.log b/wandb/run-20250916_154446-2dn97uuu/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..18a5f9aabdb7d01c1ff3b720315423a3e61848ee --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:44:46.234953827+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp87zchrlj/port-3326401.txt","pid":3326401,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:44:46.235203025+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:44:46.235870412+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3326401} +{"time":"2025-09-16T15:44:46.23584121+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3326401-3327182-1004706858/socket","Net":"unix"}} +{"time":"2025-09-16T15:44:46.42368474+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:44:46.429699161+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"2dn97uuu","id":"1(@)"} +{"time":"2025-09-16T15:44:47.146504875+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"2dn97uuu","id":"1(@)"} +{"time":"2025-09-16T15:46:01.622628196+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:46:01.622699486+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:46:01.622693814+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:46:01.622858777+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3326401-3327182-1004706858/socket","Net":"unix"}} +{"time":"2025-09-16T15:46:01.622875591+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:46:02.785870456+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:46:02.78592736+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:46:02.785957759+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_154446-2dn97uuu/logs/debug-internal.log b/wandb/run-20250916_154446-2dn97uuu/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a76c3367f8911a0228226250d879e92559b8866f --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:44:46.429890502+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:44:46.461616698+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:44:47.146352939+08:00","level":"INFO","msg":"stream: created new stream","id":"2dn97uuu"} +{"time":"2025-09-16T15:44:47.146489255+08:00","level":"INFO","msg":"stream: started","id":"2dn97uuu"} +{"time":"2025-09-16T15:44:47.146525945+08:00","level":"INFO","msg":"writer: started","stream_id":"2dn97uuu"} +{"time":"2025-09-16T15:44:47.146541939+08:00","level":"INFO","msg":"handler: started","stream_id":"2dn97uuu"} +{"time":"2025-09-16T15:44:47.146599148+08:00","level":"INFO","msg":"sender: started","stream_id":"2dn97uuu"} +{"time":"2025-09-16T15:46:01.622687082+08:00","level":"INFO","msg":"stream: closing","id":"2dn97uuu"} +{"time":"2025-09-16T15:46:02.371668887+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:46:02.784409295+08:00","level":"INFO","msg":"handler: closed","stream_id":"2dn97uuu"} +{"time":"2025-09-16T15:46:02.785261515+08:00","level":"INFO","msg":"sender: closed","stream_id":"2dn97uuu"} +{"time":"2025-09-16T15:46:02.785330833+08:00","level":"INFO","msg":"stream: closed","id":"2dn97uuu"} diff --git a/wandb/run-20250916_154446-2dn97uuu/logs/debug.log b/wandb/run-20250916_154446-2dn97uuu/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..89841b769fa6d8556160dcdc2a0bf97ec104fd86 --- /dev/null +++ b/wandb/run-20250916_154446-2dn97uuu/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_setup.py:_flush():81] Configure stats pid to 3326401 +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_154446-2dn97uuu/logs/debug.log +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_154446-2dn97uuu/logs/debug-internal.log +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:44:46,211 INFO MainThread:3326401 [wandb_init.py:init():854] starting backend +2025-09-16 15:44:46,423 INFO MainThread:3326401 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:44:46,426 INFO MainThread:3326401 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:44:46,427 INFO MainThread:3326401 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:44:46,434 INFO MainThread:3326401 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:44:47,664 INFO MainThread:3326401 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:44:47,806 INFO MainThread:3326401 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:44:47,807 INFO MainThread:3326401 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:44:47,807 INFO MainThread:3326401 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:44:47,807 INFO MainThread:3326401 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:44:47,810 INFO MainThread:3326401 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:45:56,474 INFO MainThread:3326401 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:46:01,623 INFO wandb-AsyncioManager-main:3326401 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:46:01,623 INFO wandb-AsyncioManager-main:3326401 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_154446-2dn97uuu/run-2dn97uuu.wandb b/wandb/run-20250916_154446-2dn97uuu/run-2dn97uuu.wandb new file mode 100644 index 0000000000000000000000000000000000000000..b524338a57448def837343796ff98879dcee9944 Binary files /dev/null and b/wandb/run-20250916_154446-2dn97uuu/run-2dn97uuu.wandb differ diff --git a/wandb/run-20250916_154628-nh2qmlov/files/output.log b/wandb/run-20250916_154628-nh2qmlov/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..a1ea42636dde620c7ff7520e804f998f4855d22e --- /dev/null +++ b/wandb/run-20250916_154628-nh2qmlov/files/output.log @@ -0,0 +1,8 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_154628-nh2qmlov/files/requirements.txt b/wandb/run-20250916_154628-nh2qmlov/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_154628-nh2qmlov/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_154628-nh2qmlov/files/wandb-metadata.json b/wandb/run-20250916_154628-nh2qmlov/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8c70b28b05b4c8aaac4533239620784b652a3f6e --- /dev/null +++ b/wandb/run-20250916_154628-nh2qmlov/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:46:28.185935Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577052209152" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "9qhhs3fguh98fk22084hk6zab2iqac3t" +} \ No newline at end of file diff --git a/wandb/run-20250916_154628-nh2qmlov/logs/debug-core.log b/wandb/run-20250916_154628-nh2qmlov/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..346ddaa835f1e10a4b0d4a3057f258735920215c --- /dev/null +++ b/wandb/run-20250916_154628-nh2qmlov/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:46:28.220623843+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpr9jwh3l2/port-3331216.txt","pid":3331216,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:46:28.220840218+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:46:28.221379175+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3331216} +{"time":"2025-09-16T15:46:28.22138691+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3331216-3332743-3726783727/socket","Net":"unix"}} +{"time":"2025-09-16T15:46:28.409619692+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:46:28.419224449+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"nh2qmlov","id":"1(@)"} +{"time":"2025-09-16T15:46:28.908206094+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"nh2qmlov","id":"1(@)"} +{"time":"2025-09-16T15:47:22.377810661+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_154628-nh2qmlov/logs/debug-internal.log b/wandb/run-20250916_154628-nh2qmlov/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2788c0213c2b3bae359e04cc7e37ba403c17301b --- /dev/null +++ b/wandb/run-20250916_154628-nh2qmlov/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:46:28.419484408+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:46:28.451245757+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:46:28.908109338+08:00","level":"INFO","msg":"stream: created new stream","id":"nh2qmlov"} +{"time":"2025-09-16T15:46:28.908181042+08:00","level":"INFO","msg":"stream: started","id":"nh2qmlov"} +{"time":"2025-09-16T15:46:28.908225877+08:00","level":"INFO","msg":"writer: started","stream_id":"nh2qmlov"} +{"time":"2025-09-16T15:46:28.908232078+08:00","level":"INFO","msg":"sender: started","stream_id":"nh2qmlov"} +{"time":"2025-09-16T15:46:28.908310679+08:00","level":"INFO","msg":"handler: started","stream_id":"nh2qmlov"} diff --git a/wandb/run-20250916_154628-nh2qmlov/logs/debug.log b/wandb/run-20250916_154628-nh2qmlov/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..3fda726eff306f2d4d14aab617826df78ac1fa66 --- /dev/null +++ b/wandb/run-20250916_154628-nh2qmlov/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-16 15:46:28,189 INFO MainThread:3331216 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_setup.py:_flush():81] Configure stats pid to 3331216 +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_154628-nh2qmlov/logs/debug.log +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_154628-nh2qmlov/logs/debug-internal.log +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:46:28,190 INFO MainThread:3331216 [wandb_init.py:init():854] starting backend +2025-09-16 15:46:28,409 INFO MainThread:3331216 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:46:28,414 INFO MainThread:3331216 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:46:28,417 INFO MainThread:3331216 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:46:28,427 INFO MainThread:3331216 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:46:29,530 INFO MainThread:3331216 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:46:29,717 INFO MainThread:3331216 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:46:29,718 INFO MainThread:3331216 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:46:29,718 INFO MainThread:3331216 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:46:29,718 INFO MainThread:3331216 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:46:29,721 INFO MainThread:3331216 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250916_154628-nh2qmlov/run-nh2qmlov.wandb b/wandb/run-20250916_154628-nh2qmlov/run-nh2qmlov.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_154751-py90ca1t/files/config.yaml b/wandb/run-20250916_154751-py90ca1t/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..95204f2bfcf4ada38b6338dc123fa46dd609c2e0 --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 3i5c2qdybbsv50k6bw0lrcctbo0dwprd: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577052442624" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T07:47:51.264825Z" + writerId: 3i5c2qdybbsv50k6bw0lrcctbo0dwprd + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_154751-py90ca1t/files/output.log b/wandb/run-20250916_154751-py90ca1t/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..29b46224dca2be5ca2c7496f516c2775cafb0b0c --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/files/output.log @@ -0,0 +1,137 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 538, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 616, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 620, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 160, in get_logsum_prob + log_probs = F.log_softmax(logits, dim=-1) +NameError: name 'F' is not defined +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 538, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 616, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 620, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 160, in get_logsum_prob +[rank0]: log_probs = F.log_softmax(logits, dim=-1) +[rank0]: NameError: name 'F' is not defined diff --git a/wandb/run-20250916_154751-py90ca1t/files/requirements.txt b/wandb/run-20250916_154751-py90ca1t/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_154751-py90ca1t/files/wandb-metadata.json b/wandb/run-20250916_154751-py90ca1t/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..12e77047f8d11e0c03359f9ca9b2a7b4b3c8a8ca --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:47:51.264825Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577052442624" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "3i5c2qdybbsv50k6bw0lrcctbo0dwprd" +} \ No newline at end of file diff --git a/wandb/run-20250916_154751-py90ca1t/files/wandb-summary.json b/wandb/run-20250916_154751-py90ca1t/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..bb02c2194278226895528365e676d950e985bc07 --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":89},"_runtime":89} \ No newline at end of file diff --git a/wandb/run-20250916_154751-py90ca1t/logs/debug-core.log b/wandb/run-20250916_154751-py90ca1t/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..ce70b407ee7dfe9912daaf419d09fb75aaaf51d6 --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T15:47:51.299845736+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp0a62f8pb/port-3335750.txt","pid":3335750,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:47:51.300167824+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:47:51.301139951+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3335750} +{"time":"2025-09-16T15:47:51.301131305+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3335750-3336415-1584049810/socket","Net":"unix"}} +{"time":"2025-09-16T15:47:51.489576712+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:47:51.50163765+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"py90ca1t","id":"1(@)"} +{"time":"2025-09-16T15:47:51.98199552+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"py90ca1t","id":"1(@)"} +{"time":"2025-09-16T15:49:22.163646227+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T15:49:22.164577198+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T15:49:22.164609366+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T15:49:22.164665646+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T15:49:22.164863255+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3335750-3336415-1584049810/socket","Net":"unix"}} +{"time":"2025-09-16T15:49:23.251694495+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T15:49:23.251733626+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T15:49:23.251751732+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_154751-py90ca1t/logs/debug-internal.log b/wandb/run-20250916_154751-py90ca1t/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d6578165b4728290b9d4e8081056979d8300a925 --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T15:47:51.501855132+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:47:51.529082957+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:47:51.981875871+08:00","level":"INFO","msg":"stream: created new stream","id":"py90ca1t"} +{"time":"2025-09-16T15:47:51.98198092+08:00","level":"INFO","msg":"stream: started","id":"py90ca1t"} +{"time":"2025-09-16T15:47:51.982012741+08:00","level":"INFO","msg":"writer: started","stream_id":"py90ca1t"} +{"time":"2025-09-16T15:47:51.982241977+08:00","level":"INFO","msg":"handler: started","stream_id":"py90ca1t"} +{"time":"2025-09-16T15:47:51.98226622+08:00","level":"INFO","msg":"sender: started","stream_id":"py90ca1t"} +{"time":"2025-09-16T15:49:22.164597981+08:00","level":"INFO","msg":"stream: closing","id":"py90ca1t"} +{"time":"2025-09-16T15:49:22.946500331+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T15:49:23.249716675+08:00","level":"INFO","msg":"handler: closed","stream_id":"py90ca1t"} +{"time":"2025-09-16T15:49:23.250443337+08:00","level":"INFO","msg":"sender: closed","stream_id":"py90ca1t"} +{"time":"2025-09-16T15:49:23.25050744+08:00","level":"INFO","msg":"stream: closed","id":"py90ca1t"} diff --git a/wandb/run-20250916_154751-py90ca1t/logs/debug.log b/wandb/run-20250916_154751-py90ca1t/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a721cd5bca8219c2c5376dd24d8c6277bc29151d --- /dev/null +++ b/wandb/run-20250916_154751-py90ca1t/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 15:47:51,268 INFO MainThread:3335750 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_setup.py:_flush():81] Configure stats pid to 3335750 +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_154751-py90ca1t/logs/debug.log +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_154751-py90ca1t/logs/debug-internal.log +2025-09-16 15:47:51,269 INFO MainThread:3335750 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:47:51,270 INFO MainThread:3335750 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:47:51,270 INFO MainThread:3335750 [wandb_init.py:init():854] starting backend +2025-09-16 15:47:51,489 INFO MainThread:3335750 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:47:51,494 INFO MainThread:3335750 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:47:51,498 INFO MainThread:3335750 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:47:51,512 INFO MainThread:3335750 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:47:52,318 INFO MainThread:3335750 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:47:52,495 INFO MainThread:3335750 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:47:52,495 INFO MainThread:3335750 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:47:52,495 INFO MainThread:3335750 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:47:52,495 INFO MainThread:3335750 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:47:52,498 INFO MainThread:3335750 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 15:49:00,823 INFO MainThread:3335750 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 15:49:22,162 INFO wandb-AsyncioManager-main:3335750 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 15:49:22,162 INFO wandb-AsyncioManager-main:3335750 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_154751-py90ca1t/run-py90ca1t.wandb b/wandb/run-20250916_154751-py90ca1t/run-py90ca1t.wandb new file mode 100644 index 0000000000000000000000000000000000000000..cf64b4254efce7e11c18494812091832920486db Binary files /dev/null and b/wandb/run-20250916_154751-py90ca1t/run-py90ca1t.wandb differ diff --git a/wandb/run-20250916_154950-9g1dv5jr/files/output.log b/wandb/run-20250916_154950-9g1dv5jr/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..a1ea42636dde620c7ff7520e804f998f4855d22e --- /dev/null +++ b/wandb/run-20250916_154950-9g1dv5jr/files/output.log @@ -0,0 +1,8 @@ +The length of the dataset is: 27468 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_154950-9g1dv5jr/files/requirements.txt b/wandb/run-20250916_154950-9g1dv5jr/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_154950-9g1dv5jr/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_154950-9g1dv5jr/files/wandb-metadata.json b/wandb/run-20250916_154950-9g1dv5jr/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..461fd2067059d98efc12f1095a504fa9d94dd7de --- /dev/null +++ b/wandb/run-20250916_154950-9g1dv5jr/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T07:49:50.397966Z", + "args": [ + "fit", + "--data=StrategicDialogue", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_strategic/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=Strategic-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_strategic", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577052753920" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "mdfe4ynvt9dtogkorq6befc69zooyoj3" +} \ No newline at end of file diff --git a/wandb/run-20250916_154950-9g1dv5jr/logs/debug-core.log b/wandb/run-20250916_154950-9g1dv5jr/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bfe04f1482abadcd8ff69773d31bac487db620ad --- /dev/null +++ b/wandb/run-20250916_154950-9g1dv5jr/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T15:49:50.432809164+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpmf13d8e4/port-3341509.txt","pid":3341509,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T15:49:50.433010672+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T15:49:50.435150464+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3341509} +{"time":"2025-09-16T15:49:50.435230873+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3341509-3342540-500195544/socket","Net":"unix"}} +{"time":"2025-09-16T15:49:50.621563724+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T15:49:50.632225506+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"9g1dv5jr","id":"1(@)"} +{"time":"2025-09-16T15:49:51.120711319+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"9g1dv5jr","id":"1(@)"} +{"time":"2025-09-16T15:50:42.801028068+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_154950-9g1dv5jr/logs/debug-internal.log b/wandb/run-20250916_154950-9g1dv5jr/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..135056205056912081dd3e5dc4f696379b29f561 --- /dev/null +++ b/wandb/run-20250916_154950-9g1dv5jr/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T15:49:50.632416857+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T15:49:50.658038739+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T15:49:51.120620996+08:00","level":"INFO","msg":"stream: created new stream","id":"9g1dv5jr"} +{"time":"2025-09-16T15:49:51.120696371+08:00","level":"INFO","msg":"stream: started","id":"9g1dv5jr"} +{"time":"2025-09-16T15:49:51.120822547+08:00","level":"INFO","msg":"writer: started","stream_id":"9g1dv5jr"} +{"time":"2025-09-16T15:49:51.120879851+08:00","level":"INFO","msg":"sender: started","stream_id":"9g1dv5jr"} +{"time":"2025-09-16T15:49:51.120841651+08:00","level":"INFO","msg":"handler: started","stream_id":"9g1dv5jr"} diff --git a/wandb/run-20250916_154950-9g1dv5jr/logs/debug.log b/wandb/run-20250916_154950-9g1dv5jr/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..9c6ca00a642359d9124d0eee2442cc6f27ae5a14 --- /dev/null +++ b/wandb/run-20250916_154950-9g1dv5jr/logs/debug.log @@ -0,0 +1,21 @@ +2025-09-16 15:49:50,401 INFO MainThread:3341509 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_setup.py:_flush():81] Configure stats pid to 3341509 +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_154950-9g1dv5jr/logs/debug.log +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_154950-9g1dv5jr/logs/debug-internal.log +2025-09-16 15:49:50,402 INFO MainThread:3341509 [wandb_init.py:init():813] calling init triggers +2025-09-16 15:49:50,403 INFO MainThread:3341509 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 15:49:50,403 INFO MainThread:3341509 [wandb_init.py:init():854] starting backend +2025-09-16 15:49:50,621 INFO MainThread:3341509 [wandb_init.py:init():857] sending inform_init request +2025-09-16 15:49:50,626 INFO MainThread:3341509 [wandb_init.py:init():865] backend started and connected +2025-09-16 15:49:50,629 INFO MainThread:3341509 [wandb_init.py:init():936] updated telemetry +2025-09-16 15:49:50,639 INFO MainThread:3341509 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 15:49:51,438 INFO MainThread:3341509 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 15:49:51,622 INFO MainThread:3341509 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 15:49:51,622 INFO MainThread:3341509 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 15:49:51,622 INFO MainThread:3341509 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 15:49:51,622 INFO MainThread:3341509 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 15:49:51,625 INFO MainThread:3341509 [wandb_init.py:init():1049] run started, returning control to user process diff --git a/wandb/run-20250916_154950-9g1dv5jr/run-9g1dv5jr.wandb b/wandb/run-20250916_154950-9g1dv5jr/run-9g1dv5jr.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_155110-6v3dz1c8/files/output.log b/wandb/run-20250916_155110-6v3dz1c8/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..da08029cf488b942cdc9535e747b26da6b7b944a --- /dev/null +++ b/wandb/run-20250916_155110-6v3dz1c8/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 543, in training_step + self.manual_backward(actor_loss) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 543, in training_step +[rank0]: self.manual_backward(actor_loss) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn diff --git a/wandb/run-20250916_160420-r668znk0/files/requirements.txt b/wandb/run-20250916_160420-r668znk0/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_160420-r668znk0/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_160420-r668znk0/files/wandb-metadata.json b/wandb/run-20250916_160420-r668znk0/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..23f0b7ee0875e57f2d06e5758511d9c30b3922e9 --- /dev/null +++ b/wandb/run-20250916_160420-r668znk0/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:04:20.416404Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577053708288" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "k1z1xluc1gah7p5rgajblrwkumzpzhas" +} \ No newline at end of file diff --git a/wandb/run-20250916_160420-r668znk0/files/wandb-summary.json b/wandb/run-20250916_160420-r668znk0/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e45dfcf4d1894817b0ac99d5bb319c84fdbbaffe --- /dev/null +++ b/wandb/run-20250916_160420-r668znk0/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":83},"_runtime":83} \ No newline at end of file diff --git a/wandb/run-20250916_160420-r668znk0/logs/debug-core.log b/wandb/run-20250916_160420-r668znk0/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..035e4d4bdb6fdbb0824b3411f3eed8634d20321e --- /dev/null +++ b/wandb/run-20250916_160420-r668znk0/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T16:04:20.452375056+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpkekd2vux/port-3377062.txt","pid":3377062,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:04:20.45255686+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:04:20.453053826+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3377062} +{"time":"2025-09-16T16:04:20.453070602+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3377062-3377771-3274280921/socket","Net":"unix"}} +{"time":"2025-09-16T16:04:20.640500882+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:04:20.649024374+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"r668znk0","id":"1(@)"} +{"time":"2025-09-16T16:04:21.155583401+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"r668znk0","id":"1(@)"} +{"time":"2025-09-16T16:05:44.769648877+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T16:05:44.769733253+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T16:05:44.76981969+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T16:05:44.769754716+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T16:05:44.77048504+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3377062-3377771-3274280921/socket","Net":"unix"}} +{"time":"2025-09-16T16:05:45.806933917+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T16:05:45.806993635+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T16:05:45.807025236+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_160420-r668znk0/logs/debug-internal.log b/wandb/run-20250916_160420-r668znk0/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b05ae00426828183eb5a25a10617f45c8c3de991 --- /dev/null +++ b/wandb/run-20250916_160420-r668znk0/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T16:04:20.649321217+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:04:20.681002541+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:04:21.155463835+08:00","level":"INFO","msg":"stream: created new stream","id":"r668znk0"} +{"time":"2025-09-16T16:04:21.155565938+08:00","level":"INFO","msg":"stream: started","id":"r668znk0"} +{"time":"2025-09-16T16:04:21.155582452+08:00","level":"INFO","msg":"sender: started","stream_id":"r668znk0"} +{"time":"2025-09-16T16:04:21.155578204+08:00","level":"INFO","msg":"handler: started","stream_id":"r668znk0"} +{"time":"2025-09-16T16:04:21.155607012+08:00","level":"INFO","msg":"writer: started","stream_id":"r668znk0"} +{"time":"2025-09-16T16:05:44.769755163+08:00","level":"INFO","msg":"stream: closing","id":"r668znk0"} +{"time":"2025-09-16T16:05:45.546953735+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T16:05:45.805056869+08:00","level":"INFO","msg":"handler: closed","stream_id":"r668znk0"} +{"time":"2025-09-16T16:05:45.805740186+08:00","level":"INFO","msg":"sender: closed","stream_id":"r668znk0"} +{"time":"2025-09-16T16:05:45.805802834+08:00","level":"INFO","msg":"stream: closed","id":"r668znk0"} diff --git a/wandb/run-20250916_160420-r668znk0/logs/debug.log b/wandb/run-20250916_160420-r668znk0/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..8a87ed0ec9246b301bc39843da2cb1f03bb9c70e --- /dev/null +++ b/wandb/run-20250916_160420-r668znk0/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 16:04:20,420 INFO MainThread:3377062 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:04:20,420 INFO MainThread:3377062 [wandb_setup.py:_flush():81] Configure stats pid to 3377062 +2025-09-16 16:04:20,420 INFO MainThread:3377062 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:04:20,420 INFO MainThread:3377062 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:04:20,420 INFO MainThread:3377062 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:04:20,420 INFO MainThread:3377062 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_160420-r668znk0/logs/debug.log +2025-09-16 16:04:20,421 INFO MainThread:3377062 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_160420-r668znk0/logs/debug-internal.log +2025-09-16 16:04:20,421 INFO MainThread:3377062 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:04:20,421 INFO MainThread:3377062 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:04:20,421 INFO MainThread:3377062 [wandb_init.py:init():854] starting backend +2025-09-16 16:04:20,640 INFO MainThread:3377062 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:04:20,645 INFO MainThread:3377062 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:04:20,648 INFO MainThread:3377062 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:04:20,658 INFO MainThread:3377062 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:04:21,534 INFO MainThread:3377062 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:04:21,684 INFO MainThread:3377062 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:04:21,685 INFO MainThread:3377062 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:04:21,685 INFO MainThread:3377062 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:04:21,685 INFO MainThread:3377062 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:04:21,688 INFO MainThread:3377062 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:05:30,822 INFO MainThread:3377062 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 16:05:44,769 INFO wandb-AsyncioManager-main:3377062 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 16:05:44,770 INFO wandb-AsyncioManager-main:3377062 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_160420-r668znk0/run-r668znk0.wandb b/wandb/run-20250916_160420-r668znk0/run-r668znk0.wandb new file mode 100644 index 0000000000000000000000000000000000000000..3e470ff25ffa76883d9d20491aad2efa803da4f7 Binary files /dev/null and b/wandb/run-20250916_160420-r668znk0/run-r668znk0.wandb differ diff --git a/wandb/run-20250916_160707-fmnp77ia/files/config.yaml b/wandb/run-20250916_160707-fmnp77ia/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5b8dedcdec38107998f5084d7e0625bd7b696a00 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + anf9n2mzuautupe3oscgrjtuzxkr9dnv: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577054003200" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T08:07:07.213817Z" + writerId: anf9n2mzuautupe3oscgrjtuzxkr9dnv + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_160707-fmnp77ia/files/output.log b/wandb/run-20250916_160707-fmnp77ia/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..cead7651bab3641282709c9cfc5b4531dbbb1141 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/files/output.log @@ -0,0 +1,175 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 527, in training_step + self.manual_backward(actor_loss) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 527, in training_step +[rank0]: self.manual_backward(actor_loss) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn diff --git a/wandb/run-20250916_160707-fmnp77ia/files/requirements.txt b/wandb/run-20250916_160707-fmnp77ia/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_160707-fmnp77ia/files/wandb-metadata.json b/wandb/run-20250916_160707-fmnp77ia/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c3d5a9d3e59bf625f4492491d93178b533e08b09 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:07:07.213817Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577054003200" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "anf9n2mzuautupe3oscgrjtuzxkr9dnv" +} \ No newline at end of file diff --git a/wandb/run-20250916_160707-fmnp77ia/files/wandb-summary.json b/wandb/run-20250916_160707-fmnp77ia/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e45dfcf4d1894817b0ac99d5bb319c84fdbbaffe --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":83},"_runtime":83} \ No newline at end of file diff --git a/wandb/run-20250916_160707-fmnp77ia/logs/debug-core.log b/wandb/run-20250916_160707-fmnp77ia/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bcb1caa06f2302dac0a5b582e78a52049d64f558 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-16T16:07:07.248242194+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpqcuda22x/port-3385304.txt","pid":3385304,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:07:07.248498927+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:07:07.249458147+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3385304} +{"time":"2025-09-16T16:07:07.249464663+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3385304-3386062-102470166/socket","Net":"unix"}} +{"time":"2025-09-16T16:07:07.438048409+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:07:07.445836568+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"fmnp77ia","id":"1(@)"} +{"time":"2025-09-16T16:07:07.928934887+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"fmnp77ia","id":"1(@)"} +{"time":"2025-09-16T16:08:31.971720821+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T16:08:31.971855697+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T16:08:31.971835274+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T16:08:31.972000066+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T16:08:31.972265091+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3385304-3386062-102470166/socket","Net":"unix"}} +{"time":"2025-09-16T16:08:34.695617329+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_160707-fmnp77ia/logs/debug-internal.log b/wandb/run-20250916_160707-fmnp77ia/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b794d6f490731539667a5fc03d9adc8d23f12f54 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T16:07:07.445962398+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:07:07.472764312+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:07:07.928822052+08:00","level":"INFO","msg":"stream: created new stream","id":"fmnp77ia"} +{"time":"2025-09-16T16:07:07.928919074+08:00","level":"INFO","msg":"stream: started","id":"fmnp77ia"} +{"time":"2025-09-16T16:07:07.928970533+08:00","level":"INFO","msg":"writer: started","stream_id":"fmnp77ia"} +{"time":"2025-09-16T16:07:07.928978715+08:00","level":"INFO","msg":"sender: started","stream_id":"fmnp77ia"} +{"time":"2025-09-16T16:07:07.929011991+08:00","level":"INFO","msg":"handler: started","stream_id":"fmnp77ia"} +{"time":"2025-09-16T16:08:31.971843924+08:00","level":"INFO","msg":"stream: closing","id":"fmnp77ia"} diff --git a/wandb/run-20250916_160707-fmnp77ia/logs/debug.log b/wandb/run-20250916_160707-fmnp77ia/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..836025978b1feaf2e4755a8c7885c73e003bda36 --- /dev/null +++ b/wandb/run-20250916_160707-fmnp77ia/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 16:07:07,217 INFO MainThread:3385304 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:07:07,217 INFO MainThread:3385304 [wandb_setup.py:_flush():81] Configure stats pid to 3385304 +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_160707-fmnp77ia/logs/debug.log +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_160707-fmnp77ia/logs/debug-internal.log +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:07:07,218 INFO MainThread:3385304 [wandb_init.py:init():854] starting backend +2025-09-16 16:07:07,438 INFO MainThread:3385304 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:07:07,443 INFO MainThread:3385304 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:07:07,445 INFO MainThread:3385304 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:07:07,456 INFO MainThread:3385304 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:07:08,249 INFO MainThread:3385304 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:07:08,428 INFO MainThread:3385304 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:07:08,430 INFO MainThread:3385304 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:07:08,430 INFO MainThread:3385304 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:07:08,430 INFO MainThread:3385304 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:07:08,433 INFO MainThread:3385304 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:08:17,403 INFO MainThread:3385304 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 16:08:31,971 INFO wandb-AsyncioManager-main:3385304 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 16:08:31,971 INFO wandb-AsyncioManager-main:3385304 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_160707-fmnp77ia/run-fmnp77ia.wandb b/wandb/run-20250916_160707-fmnp77ia/run-fmnp77ia.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_161136-74uyp8vt/files/output.log b/wandb/run-20250916_161136-74uyp8vt/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8f8c7a2c70dae10493729271436a0e0e4388d455 --- /dev/null +++ b/wandb/run-20250916_161136-74uyp8vt/files/output.log @@ -0,0 +1,11 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 1/1652 [00:24<11:08:45, 0.04it/s, v_num=p8vt] +/home/jiashuo/codes/OfflineArcher/Algorithms.py:129: FutureWarning: `torch.cuda.amp.autocast(args...)` is deprecated. Please use `torch.amp.autocast('cuda', args...)` instead. + with torch.cuda.amp.autocast(): # 添加半精度和no_grad diff --git a/wandb/run-20250916_161136-74uyp8vt/files/requirements.txt b/wandb/run-20250916_161136-74uyp8vt/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_161136-74uyp8vt/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_161136-74uyp8vt/files/wandb-metadata.json b/wandb/run-20250916_161136-74uyp8vt/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a87dc6b67658a2beeba3e57c6186f6ccc995271b --- /dev/null +++ b/wandb/run-20250916_161136-74uyp8vt/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:11:36.367845Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577054543872" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "12rj4798ppyybpwpqjhjdqnoaa3nx1e0" +} \ No newline at end of file diff --git a/wandb/run-20250916_161136-74uyp8vt/logs/debug-core.log b/wandb/run-20250916_161136-74uyp8vt/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..439314b4652039a343b3262e0a92f1bc37a73410 --- /dev/null +++ b/wandb/run-20250916_161136-74uyp8vt/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T16:11:36.403068905+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp23r3epzm/port-3396694.txt","pid":3396694,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:11:36.403289898+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:11:36.404027487+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3396694} +{"time":"2025-09-16T16:11:36.404008969+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3396694-3397920-2549052858/socket","Net":"unix"}} +{"time":"2025-09-16T16:11:36.591928648+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:11:36.599495787+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"74uyp8vt","id":"1(@)"} +{"time":"2025-09-16T16:11:37.372169947+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"74uyp8vt","id":"1(@)"} +{"time":"2025-09-16T16:15:55.228348302+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_161136-74uyp8vt/logs/debug-internal.log b/wandb/run-20250916_161136-74uyp8vt/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..ad6af10a190fbdae669a1bc21d915e05291b77ba --- /dev/null +++ b/wandb/run-20250916_161136-74uyp8vt/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T16:11:36.599674021+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:11:36.626361773+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:11:37.372057948+08:00","level":"INFO","msg":"stream: created new stream","id":"74uyp8vt"} +{"time":"2025-09-16T16:11:37.372154454+08:00","level":"INFO","msg":"stream: started","id":"74uyp8vt"} +{"time":"2025-09-16T16:11:37.372202694+08:00","level":"INFO","msg":"writer: started","stream_id":"74uyp8vt"} +{"time":"2025-09-16T16:11:37.372245555+08:00","level":"INFO","msg":"handler: started","stream_id":"74uyp8vt"} +{"time":"2025-09-16T16:11:37.372261835+08:00","level":"INFO","msg":"sender: started","stream_id":"74uyp8vt"} diff --git a/wandb/run-20250916_161136-74uyp8vt/logs/debug.log b/wandb/run-20250916_161136-74uyp8vt/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..3590bd9dca807e8fca2b36c248936678b04debf9 --- /dev/null +++ b/wandb/run-20250916_161136-74uyp8vt/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 16:11:36,371 INFO MainThread:3396694 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:11:36,371 INFO MainThread:3396694 [wandb_setup.py:_flush():81] Configure stats pid to 3396694 +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_161136-74uyp8vt/logs/debug.log +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_161136-74uyp8vt/logs/debug-internal.log +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:11:36,372 INFO MainThread:3396694 [wandb_init.py:init():854] starting backend +2025-09-16 16:11:36,592 INFO MainThread:3396694 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:11:36,596 INFO MainThread:3396694 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:11:36,599 INFO MainThread:3396694 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:11:36,609 INFO MainThread:3396694 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:11:37,697 INFO MainThread:3396694 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:11:37,847 INFO MainThread:3396694 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:11:37,847 INFO MainThread:3396694 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:11:37,847 INFO MainThread:3396694 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:11:37,847 INFO MainThread:3396694 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:11:37,850 INFO MainThread:3396694 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:12:58,236 INFO MainThread:3396694 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_161136-74uyp8vt/run-74uyp8vt.wandb b/wandb/run-20250916_161136-74uyp8vt/run-74uyp8vt.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_161442-sb673vv0/files/config.yaml b/wandb/run-20250916_161442-sb673vv0/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..97f9a92bd815d590b1d18b355ec53e7d94d9e0fa --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + hwzqjn3c9rdz56gr1egt7drz3ehtzmq7: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577054920704" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T08:14:42.730010Z" + writerId: hwzqjn3c9rdz56gr1egt7drz3ehtzmq7 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_161442-sb673vv0/files/output.log b/wandb/run-20250916_161442-sb673vv0/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_161442-sb673vv0/files/requirements.txt b/wandb/run-20250916_161442-sb673vv0/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_161442-sb673vv0/files/wandb-metadata.json b/wandb/run-20250916_161442-sb673vv0/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c4ff62a584d8280b95fb4af0d68a73e0fb995ce4 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:14:42.730010Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577054920704" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "hwzqjn3c9rdz56gr1egt7drz3ehtzmq7" +} \ No newline at end of file diff --git a/wandb/run-20250916_161442-sb673vv0/files/wandb-summary.json b/wandb/run-20250916_161442-sb673vv0/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..7b7f6bbcb78630a18aad85b8e448b0042f42c288 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":39},"_runtime":39} \ No newline at end of file diff --git a/wandb/run-20250916_161442-sb673vv0/logs/debug-core.log b/wandb/run-20250916_161442-sb673vv0/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..6daa6b5787c62b22c0f5fb862c239012c635fa46 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T16:14:42.754649195+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpbpdb1jte/port-3405962.txt","pid":3405962,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:14:42.755456805+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:14:42.757279281+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3405962} +{"time":"2025-09-16T16:14:42.757246712+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3405962-3406771-4089463540/socket","Net":"unix"}} +{"time":"2025-09-16T16:14:42.943641908+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:14:42.951349476+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"sb673vv0","id":"1(@)"} +{"time":"2025-09-16T16:14:43.428323339+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"sb673vv0","id":"1(@)"} +{"time":"2025-09-16T16:15:22.928194557+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T16:15:22.928285084+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T16:15:22.928270205+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T16:15:22.928446523+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T16:15:22.928582185+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3405962-3406771-4089463540/socket","Net":"unix"}} +{"time":"2025-09-16T16:15:23.872306029+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T16:15:23.87235072+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T16:15:23.872371667+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_161442-sb673vv0/logs/debug-internal.log b/wandb/run-20250916_161442-sb673vv0/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b004e35aa4313bff357ff8dfe17e65fd2328b162 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T16:14:42.951514569+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:14:42.973109502+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:14:43.4282578+08:00","level":"INFO","msg":"stream: created new stream","id":"sb673vv0"} +{"time":"2025-09-16T16:14:43.428313479+08:00","level":"INFO","msg":"stream: started","id":"sb673vv0"} +{"time":"2025-09-16T16:14:43.428329943+08:00","level":"INFO","msg":"writer: started","stream_id":"sb673vv0"} +{"time":"2025-09-16T16:14:43.428363723+08:00","level":"INFO","msg":"handler: started","stream_id":"sb673vv0"} +{"time":"2025-09-16T16:14:43.428403034+08:00","level":"INFO","msg":"sender: started","stream_id":"sb673vv0"} +{"time":"2025-09-16T16:15:22.928243503+08:00","level":"INFO","msg":"stream: closing","id":"sb673vv0"} +{"time":"2025-09-16T16:15:23.609900231+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T16:15:23.870631065+08:00","level":"INFO","msg":"handler: closed","stream_id":"sb673vv0"} +{"time":"2025-09-16T16:15:23.871155671+08:00","level":"INFO","msg":"sender: closed","stream_id":"sb673vv0"} +{"time":"2025-09-16T16:15:23.871227202+08:00","level":"INFO","msg":"stream: closed","id":"sb673vv0"} diff --git a/wandb/run-20250916_161442-sb673vv0/logs/debug.log b/wandb/run-20250916_161442-sb673vv0/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..330acace712fecde452f39ca65609c9d581ab878 --- /dev/null +++ b/wandb/run-20250916_161442-sb673vv0/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 16:14:42,731 INFO MainThread:3405962 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:14:42,731 INFO MainThread:3405962 [wandb_setup.py:_flush():81] Configure stats pid to 3405962 +2025-09-16 16:14:42,731 INFO MainThread:3405962 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:14:42,731 INFO MainThread:3405962 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:14:42,731 INFO MainThread:3405962 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:14:42,732 INFO MainThread:3405962 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_161442-sb673vv0/logs/debug.log +2025-09-16 16:14:42,732 INFO MainThread:3405962 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_161442-sb673vv0/logs/debug-internal.log +2025-09-16 16:14:42,732 INFO MainThread:3405962 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:14:42,732 INFO MainThread:3405962 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:14:42,732 INFO MainThread:3405962 [wandb_init.py:init():854] starting backend +2025-09-16 16:14:42,943 INFO MainThread:3405962 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:14:42,945 INFO MainThread:3405962 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:14:42,946 INFO MainThread:3405962 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:14:42,953 INFO MainThread:3405962 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:14:43,692 INFO MainThread:3405962 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:14:43,837 INFO MainThread:3405962 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:14:43,837 INFO MainThread:3405962 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:14:43,837 INFO MainThread:3405962 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:14:43,837 INFO MainThread:3405962 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:14:43,840 INFO MainThread:3405962 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:15:22,928 INFO wandb-AsyncioManager-main:3405962 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 16:15:22,928 INFO wandb-AsyncioManager-main:3405962 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_161442-sb673vv0/run-sb673vv0.wandb b/wandb/run-20250916_161442-sb673vv0/run-sb673vv0.wandb new file mode 100644 index 0000000000000000000000000000000000000000..522b3b7a2f26d60a8a9b9cfc439624abe1b29a31 Binary files /dev/null and b/wandb/run-20250916_161442-sb673vv0/run-sb673vv0.wandb differ diff --git a/wandb/run-20250916_161814-o286q0li/files/config.yaml b/wandb/run-20250916_161814-o286q0li/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d63a017bd8630b24c47a4975af93072fd21cdc42 --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 4opeol69u0m8hw5yyeyu0u80bedi9pll: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577055166464" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T08:18:14.421470Z" + writerId: 4opeol69u0m8hw5yyeyu0u80bedi9pll + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_161814-o286q0li/files/output.log b/wandb/run-20250916_161814-o286q0li/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..215b8a908ef48ba4e45dcd3b895ab0cd4df9ba12 --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/files/output.log @@ -0,0 +1,173 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 543, in training_step + self.manual_backward(actor_loss) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 543, in training_step +[rank0]: self.manual_backward(actor_loss) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn diff --git a/wandb/run-20250916_161814-o286q0li/files/requirements.txt b/wandb/run-20250916_161814-o286q0li/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_161814-o286q0li/files/wandb-metadata.json b/wandb/run-20250916_161814-o286q0li/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3fec1b8fc14eac3ca8f5a6f3b61fbae4f172859c --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:18:14.421470Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577055166464" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "4opeol69u0m8hw5yyeyu0u80bedi9pll" +} \ No newline at end of file diff --git a/wandb/run-20250916_161814-o286q0li/files/wandb-summary.json b/wandb/run-20250916_161814-o286q0li/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..dcdfe068b7fce1b91b105a1cf2d7aa757e7d7661 --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":82},"_runtime":82} \ No newline at end of file diff --git a/wandb/run-20250916_161814-o286q0li/logs/debug-core.log b/wandb/run-20250916_161814-o286q0li/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..3491bd572d873202386d04167ab72d0079390323 --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T16:18:14.473064586+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpipqse33s/port-3414965.txt","pid":3414965,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:18:14.473386529+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:18:14.474579915+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3414965} +{"time":"2025-09-16T16:18:14.474526976+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3414965-3415656-3749531936/socket","Net":"unix"}} +{"time":"2025-09-16T16:18:14.658525795+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:18:14.672139347+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"o286q0li","id":"1(@)"} +{"time":"2025-09-16T16:18:15.131016228+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"o286q0li","id":"1(@)"} +{"time":"2025-09-16T16:19:38.365321381+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T16:19:38.365438553+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T16:19:38.365464772+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T16:19:38.365550569+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T16:19:38.365699454+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3414965-3415656-3749531936/socket","Net":"unix"}} +{"time":"2025-09-16T16:19:40.206092518+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T16:19:40.206127301+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T16:19:40.206142629+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_161814-o286q0li/logs/debug-internal.log b/wandb/run-20250916_161814-o286q0li/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..10d5e17bbb6a48cef53d3778f2e74aad8acc4071 --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T16:18:14.672389265+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:18:14.690392212+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:18:15.130836852+08:00","level":"INFO","msg":"stream: created new stream","id":"o286q0li"} +{"time":"2025-09-16T16:18:15.130999687+08:00","level":"INFO","msg":"stream: started","id":"o286q0li"} +{"time":"2025-09-16T16:18:15.131025824+08:00","level":"INFO","msg":"sender: started","stream_id":"o286q0li"} +{"time":"2025-09-16T16:18:15.131023351+08:00","level":"INFO","msg":"handler: started","stream_id":"o286q0li"} +{"time":"2025-09-16T16:18:15.131063732+08:00","level":"INFO","msg":"writer: started","stream_id":"o286q0li"} +{"time":"2025-09-16T16:19:38.365462053+08:00","level":"INFO","msg":"stream: closing","id":"o286q0li"} +{"time":"2025-09-16T16:19:39.694585394+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T16:19:40.204444249+08:00","level":"INFO","msg":"handler: closed","stream_id":"o286q0li"} +{"time":"2025-09-16T16:19:40.205003594+08:00","level":"INFO","msg":"sender: closed","stream_id":"o286q0li"} +{"time":"2025-09-16T16:19:40.205025259+08:00","level":"INFO","msg":"stream: closed","id":"o286q0li"} diff --git a/wandb/run-20250916_161814-o286q0li/logs/debug.log b/wandb/run-20250916_161814-o286q0li/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..53c35282f47a2f1524f18b2f8687667c8adc22ec --- /dev/null +++ b/wandb/run-20250916_161814-o286q0li/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_setup.py:_flush():81] Configure stats pid to 3414965 +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_161814-o286q0li/logs/debug.log +2025-09-16 16:18:14,427 INFO MainThread:3414965 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_161814-o286q0li/logs/debug-internal.log +2025-09-16 16:18:14,428 INFO MainThread:3414965 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:18:14,428 INFO MainThread:3414965 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:18:14,428 INFO MainThread:3414965 [wandb_init.py:init():854] starting backend +2025-09-16 16:18:14,659 INFO MainThread:3414965 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:18:14,664 INFO MainThread:3414965 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:18:14,667 INFO MainThread:3414965 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:18:14,680 INFO MainThread:3414965 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:18:15,489 INFO MainThread:3414965 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:18:15,606 INFO MainThread:3414965 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:18:15,607 INFO MainThread:3414965 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:18:15,607 INFO MainThread:3414965 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:18:15,608 INFO MainThread:3414965 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:18:15,610 INFO MainThread:3414965 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:19:24,609 INFO MainThread:3414965 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 16:19:38,366 INFO wandb-AsyncioManager-main:3414965 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 16:19:38,366 INFO wandb-AsyncioManager-main:3414965 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_161814-o286q0li/run-o286q0li.wandb b/wandb/run-20250916_161814-o286q0li/run-o286q0li.wandb new file mode 100644 index 0000000000000000000000000000000000000000..84fff167697b8d37481664a1824e8800c596d482 Binary files /dev/null and b/wandb/run-20250916_161814-o286q0li/run-o286q0li.wandb differ diff --git a/wandb/run-20250916_162228-19tctv0w/files/output.log b/wandb/run-20250916_162228-19tctv0w/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..f83e12ba677ecac1e746c3076fe3092b096fddc2 --- /dev/null +++ b/wandb/run-20250916_162228-19tctv0w/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:11:43, 0.04it/s, v_num=tv0w] diff --git a/wandb/run-20250916_162228-19tctv0w/files/requirements.txt b/wandb/run-20250916_162228-19tctv0w/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_162228-19tctv0w/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_162228-19tctv0w/files/wandb-metadata.json b/wandb/run-20250916_162228-19tctv0w/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c97e8dcdcab581db41253b03a850e04fdc4f7a25 --- /dev/null +++ b/wandb/run-20250916_162228-19tctv0w/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:22:28.477230Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577055711232" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "38btlq3ch38nsztyc9ipx4n7gq3fwf3n" +} \ No newline at end of file diff --git a/wandb/run-20250916_162228-19tctv0w/logs/debug-core.log b/wandb/run-20250916_162228-19tctv0w/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bc0d7d47c7d799deab4079103fdc5915dd67f395 --- /dev/null +++ b/wandb/run-20250916_162228-19tctv0w/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T16:22:28.512920116+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpn47re7ij/port-3426113.txt","pid":3426113,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:22:28.513121539+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:22:28.513710018+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3426113} +{"time":"2025-09-16T16:22:28.513714465+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3426113-3427405-929628457/socket","Net":"unix"}} +{"time":"2025-09-16T16:22:28.701423248+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:22:28.708365896+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"19tctv0w","id":"1(@)"} +{"time":"2025-09-16T16:22:29.192313026+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"19tctv0w","id":"1(@)"} +{"time":"2025-09-16T16:27:09.988807861+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_162228-19tctv0w/logs/debug-internal.log b/wandb/run-20250916_162228-19tctv0w/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..75c5876031c7bfcea7281bb26fb594ff1700ee5e --- /dev/null +++ b/wandb/run-20250916_162228-19tctv0w/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T16:22:28.708510732+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:22:28.733343867+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:22:29.192143121+08:00","level":"INFO","msg":"stream: created new stream","id":"19tctv0w"} +{"time":"2025-09-16T16:22:29.192296846+08:00","level":"INFO","msg":"stream: started","id":"19tctv0w"} +{"time":"2025-09-16T16:22:29.192359394+08:00","level":"INFO","msg":"writer: started","stream_id":"19tctv0w"} +{"time":"2025-09-16T16:22:29.192427672+08:00","level":"INFO","msg":"handler: started","stream_id":"19tctv0w"} +{"time":"2025-09-16T16:22:29.192452605+08:00","level":"INFO","msg":"sender: started","stream_id":"19tctv0w"} diff --git a/wandb/run-20250916_162228-19tctv0w/logs/debug.log b/wandb/run-20250916_162228-19tctv0w/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..b5c32afe919b774437ccddfa400129baab066406 --- /dev/null +++ b/wandb/run-20250916_162228-19tctv0w/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_setup.py:_flush():81] Configure stats pid to 3426113 +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_162228-19tctv0w/logs/debug.log +2025-09-16 16:22:28,481 INFO MainThread:3426113 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_162228-19tctv0w/logs/debug-internal.log +2025-09-16 16:22:28,482 INFO MainThread:3426113 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:22:28,482 INFO MainThread:3426113 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:22:28,482 INFO MainThread:3426113 [wandb_init.py:init():854] starting backend +2025-09-16 16:22:28,701 INFO MainThread:3426113 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:22:28,706 INFO MainThread:3426113 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:22:28,709 INFO MainThread:3426113 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:22:28,719 INFO MainThread:3426113 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:22:29,544 INFO MainThread:3426113 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:22:29,710 INFO MainThread:3426113 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:22:29,710 INFO MainThread:3426113 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:22:29,711 INFO MainThread:3426113 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:22:29,711 INFO MainThread:3426113 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:22:29,715 INFO MainThread:3426113 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:23:38,951 INFO MainThread:3426113 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_162228-19tctv0w/run-19tctv0w.wandb b/wandb/run-20250916_162228-19tctv0w/run-19tctv0w.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_163244-dxqrxz25/files/output.log b/wandb/run-20250916_163244-dxqrxz25/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8fc2d9b2b85c3ad7492e7a4c33246e961190b9c5 --- /dev/null +++ b/wandb/run-20250916_163244-dxqrxz25/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:11:11, 0.04it/s, v_num=xz25] diff --git a/wandb/run-20250916_163244-dxqrxz25/files/requirements.txt b/wandb/run-20250916_163244-dxqrxz25/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_163244-dxqrxz25/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_163244-dxqrxz25/files/wandb-metadata.json b/wandb/run-20250916_163244-dxqrxz25/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..f846e24c7b87bba88000711b95295db282428d12 --- /dev/null +++ b/wandb/run-20250916_163244-dxqrxz25/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:32:44.656964Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577056366592" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "9h1unyrbozuxa1jwxpxgb3cy60yy3atw" +} \ No newline at end of file diff --git a/wandb/run-20250916_163244-dxqrxz25/logs/debug-core.log b/wandb/run-20250916_163244-dxqrxz25/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..826cc0f806ec1b8a03eea3b6adf29435b3927114 --- /dev/null +++ b/wandb/run-20250916_163244-dxqrxz25/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T16:32:44.691765182+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpvdshn42k/port-3451998.txt","pid":3451998,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:32:44.692021189+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:32:44.693591217+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3451998-3452782-2733519145/socket","Net":"unix"}} +{"time":"2025-09-16T16:32:44.693714428+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3451998} +{"time":"2025-09-16T16:32:44.88189585+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:32:44.889781231+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"dxqrxz25","id":"1(@)"} +{"time":"2025-09-16T16:32:45.35420807+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"dxqrxz25","id":"1(@)"} +{"time":"2025-09-16T16:35:31.953479968+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_163244-dxqrxz25/logs/debug-internal.log b/wandb/run-20250916_163244-dxqrxz25/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..fe3b35247a09fa462ad779ccac401ad0b87f96ad --- /dev/null +++ b/wandb/run-20250916_163244-dxqrxz25/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T16:32:44.889948163+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:32:44.915250177+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:32:45.354080718+08:00","level":"INFO","msg":"stream: created new stream","id":"dxqrxz25"} +{"time":"2025-09-16T16:32:45.354175261+08:00","level":"INFO","msg":"stream: started","id":"dxqrxz25"} +{"time":"2025-09-16T16:32:45.35421929+08:00","level":"INFO","msg":"writer: started","stream_id":"dxqrxz25"} +{"time":"2025-09-16T16:32:45.35423198+08:00","level":"INFO","msg":"handler: started","stream_id":"dxqrxz25"} +{"time":"2025-09-16T16:32:45.354342155+08:00","level":"INFO","msg":"sender: started","stream_id":"dxqrxz25"} diff --git a/wandb/run-20250916_163244-dxqrxz25/logs/debug.log b/wandb/run-20250916_163244-dxqrxz25/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..350cdf2184deeb8fdc8ccdd05ceedcb3f4b737cd --- /dev/null +++ b/wandb/run-20250916_163244-dxqrxz25/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 16:32:44,660 INFO MainThread:3451998 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_setup.py:_flush():81] Configure stats pid to 3451998 +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_163244-dxqrxz25/logs/debug.log +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_163244-dxqrxz25/logs/debug-internal.log +2025-09-16 16:32:44,661 INFO MainThread:3451998 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:32:44,662 INFO MainThread:3451998 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:32:44,662 INFO MainThread:3451998 [wandb_init.py:init():854] starting backend +2025-09-16 16:32:44,882 INFO MainThread:3451998 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:32:44,886 INFO MainThread:3451998 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:32:44,889 INFO MainThread:3451998 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:32:44,900 INFO MainThread:3451998 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:32:45,713 INFO MainThread:3451998 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:32:45,892 INFO MainThread:3451998 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:32:45,893 INFO MainThread:3451998 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:32:45,893 INFO MainThread:3451998 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:32:45,893 INFO MainThread:3451998 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:32:45,896 INFO MainThread:3451998 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:33:55,092 INFO MainThread:3451998 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_163244-dxqrxz25/run-dxqrxz25.wandb b/wandb/run-20250916_163244-dxqrxz25/run-dxqrxz25.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_163831-hy6iid1z/files/config.yaml b/wandb/run-20250916_163831-hy6iid1z/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..11a7bf130f94f8ee8beeffb7c06849dc26b350b6 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + ci4wkzb5xjctzjo31pip83gsufmaq1dw: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577056915456" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T08:38:31.027055Z" + writerId: ci4wkzb5xjctzjo31pip83gsufmaq1dw + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_163831-hy6iid1z/files/output.log b/wandb/run-20250916_163831-hy6iid1z/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..35c8803740a84e2392de919e5c218c7366e82fd1 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/files/output.log @@ -0,0 +1,175 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 537, in training_step + self.manual_backward(actor_loss) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 537, in training_step +[rank0]: self.manual_backward(actor_loss) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn diff --git a/wandb/run-20250916_163831-hy6iid1z/files/requirements.txt b/wandb/run-20250916_163831-hy6iid1z/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_163831-hy6iid1z/files/wandb-metadata.json b/wandb/run-20250916_163831-hy6iid1z/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3177ced3ff22114c89e98169909498f12b89d041 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:38:31.027055Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577056915456" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "ci4wkzb5xjctzjo31pip83gsufmaq1dw" +} \ No newline at end of file diff --git a/wandb/run-20250916_163831-hy6iid1z/files/wandb-summary.json b/wandb/run-20250916_163831-hy6iid1z/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e91be8747df415748ea449391fafd0573d532298 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":84},"_runtime":84} \ No newline at end of file diff --git a/wandb/run-20250916_163831-hy6iid1z/logs/debug-core.log b/wandb/run-20250916_163831-hy6iid1z/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..327fbe9be344571760f17ce3201dc42a42a27023 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T16:38:31.062386415+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpgm44usjl/port-3466233.txt","pid":3466233,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:38:31.062643451+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:38:31.063327413+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3466233} +{"time":"2025-09-16T16:38:31.063307083+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3466233-3467544-480850454/socket","Net":"unix"}} +{"time":"2025-09-16T16:38:31.251830062+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:38:31.260051958+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"hy6iid1z","id":"1(@)"} +{"time":"2025-09-16T16:38:31.725800631+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"hy6iid1z","id":"1(@)"} +{"time":"2025-09-16T16:39:56.50944688+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T16:39:56.509598251+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T16:39:56.509580854+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T16:39:56.509750342+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T16:39:56.509926984+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3466233-3467544-480850454/socket","Net":"unix"}} +{"time":"2025-09-16T16:39:58.147501561+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T16:39:58.147560784+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T16:39:58.147594839+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_163831-hy6iid1z/logs/debug-internal.log b/wandb/run-20250916_163831-hy6iid1z/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..23a496bb3f4c64864c22ca35267d321341001b38 --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T16:38:31.260213201+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:38:31.275470902+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:38:31.725647137+08:00","level":"INFO","msg":"stream: created new stream","id":"hy6iid1z"} +{"time":"2025-09-16T16:38:31.725784246+08:00","level":"INFO","msg":"stream: started","id":"hy6iid1z"} +{"time":"2025-09-16T16:38:31.725815291+08:00","level":"INFO","msg":"sender: started","stream_id":"hy6iid1z"} +{"time":"2025-09-16T16:38:31.72582355+08:00","level":"INFO","msg":"writer: started","stream_id":"hy6iid1z"} +{"time":"2025-09-16T16:38:31.72586801+08:00","level":"INFO","msg":"handler: started","stream_id":"hy6iid1z"} +{"time":"2025-09-16T16:39:56.509603987+08:00","level":"INFO","msg":"stream: closing","id":"hy6iid1z"} +{"time":"2025-09-16T16:39:57.83112986+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T16:39:58.145494432+08:00","level":"INFO","msg":"handler: closed","stream_id":"hy6iid1z"} +{"time":"2025-09-16T16:39:58.146210991+08:00","level":"INFO","msg":"sender: closed","stream_id":"hy6iid1z"} +{"time":"2025-09-16T16:39:58.146290712+08:00","level":"INFO","msg":"stream: closed","id":"hy6iid1z"} diff --git a/wandb/run-20250916_163831-hy6iid1z/logs/debug.log b/wandb/run-20250916_163831-hy6iid1z/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..faf2a1fce4263153130d371dea4e60379dfa0b3c --- /dev/null +++ b/wandb/run-20250916_163831-hy6iid1z/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 16:38:31,031 INFO MainThread:3466233 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:38:31,031 INFO MainThread:3466233 [wandb_setup.py:_flush():81] Configure stats pid to 3466233 +2025-09-16 16:38:31,031 INFO MainThread:3466233 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:38:31,031 INFO MainThread:3466233 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:38:31,031 INFO MainThread:3466233 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:38:31,031 INFO MainThread:3466233 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_163831-hy6iid1z/logs/debug.log +2025-09-16 16:38:31,032 INFO MainThread:3466233 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_163831-hy6iid1z/logs/debug-internal.log +2025-09-16 16:38:31,032 INFO MainThread:3466233 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:38:31,032 INFO MainThread:3466233 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:38:31,032 INFO MainThread:3466233 [wandb_init.py:init():854] starting backend +2025-09-16 16:38:31,252 INFO MainThread:3466233 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:38:31,256 INFO MainThread:3466233 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:38:31,259 INFO MainThread:3466233 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:38:31,270 INFO MainThread:3466233 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:38:32,046 INFO MainThread:3466233 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:38:32,213 INFO MainThread:3466233 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:38:32,214 INFO MainThread:3466233 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:38:32,214 INFO MainThread:3466233 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:38:32,214 INFO MainThread:3466233 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:38:32,217 INFO MainThread:3466233 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:39:41,421 INFO MainThread:3466233 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 16:39:56,509 INFO wandb-AsyncioManager-main:3466233 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 16:39:56,510 INFO wandb-AsyncioManager-main:3466233 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_163831-hy6iid1z/run-hy6iid1z.wandb b/wandb/run-20250916_163831-hy6iid1z/run-hy6iid1z.wandb new file mode 100644 index 0000000000000000000000000000000000000000..81f0a27ea93659cb829f52aa414a33cb9b73221f Binary files /dev/null and b/wandb/run-20250916_163831-hy6iid1z/run-hy6iid1z.wandb differ diff --git a/wandb/run-20250916_164125-tfqi7r4k/files/config.yaml b/wandb/run-20250916_164125-tfqi7r4k/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5b782bf84d6ad63b11e0d918e7fa7c9c8b384ff2 --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + kpyuvb6t0txepg411hw2tuusbtknwh7j: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577057366016" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T08:41:25.548290Z" + writerId: kpyuvb6t0txepg411hw2tuusbtknwh7j + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_164125-tfqi7r4k/files/output.log b/wandb/run-20250916_164125-tfqi7r4k/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..35c8803740a84e2392de919e5c218c7366e82fd1 --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/files/output.log @@ -0,0 +1,175 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 537, in training_step + self.manual_backward(actor_loss) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 537, in training_step +[rank0]: self.manual_backward(actor_loss) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn diff --git a/wandb/run-20250916_164125-tfqi7r4k/files/requirements.txt b/wandb/run-20250916_164125-tfqi7r4k/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_164125-tfqi7r4k/files/wandb-metadata.json b/wandb/run-20250916_164125-tfqi7r4k/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..874b5164cc8369a6a40fd60e6263db8766fce242 --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:41:25.548290Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577057366016" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "kpyuvb6t0txepg411hw2tuusbtknwh7j" +} \ No newline at end of file diff --git a/wandb/run-20250916_164125-tfqi7r4k/files/wandb-summary.json b/wandb/run-20250916_164125-tfqi7r4k/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..2d0639155343dcebf02dc833d4e9f998d2b980a2 --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":83,"_wandb":{"runtime":83}} \ No newline at end of file diff --git a/wandb/run-20250916_164125-tfqi7r4k/logs/debug-core.log b/wandb/run-20250916_164125-tfqi7r4k/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..9429cefaef5f953164274a27cfc5a4d6ddff2a3b --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T16:41:25.583880664+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp5cfsbexr/port-3474582.txt","pid":3474582,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:41:25.584116519+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:41:25.58468871+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3474582} +{"time":"2025-09-16T16:41:25.584656576+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3474582-3475479-3638167802/socket","Net":"unix"}} +{"time":"2025-09-16T16:41:25.771922781+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:41:25.782680734+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"tfqi7r4k","id":"1(@)"} +{"time":"2025-09-16T16:41:26.244921685+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"tfqi7r4k","id":"1(@)"} +{"time":"2025-09-16T16:42:49.838803751+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T16:42:49.838870818+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T16:42:49.838908003+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T16:42:49.838973119+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T16:42:49.839125302+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3474582-3475479-3638167802/socket","Net":"unix"}} +{"time":"2025-09-16T16:42:51.437379447+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T16:42:51.437449045+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T16:42:51.437489204+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_164125-tfqi7r4k/logs/debug-internal.log b/wandb/run-20250916_164125-tfqi7r4k/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e2531aa95139959c97753413f340d5d7a89526e6 --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T16:41:25.782944991+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:41:25.812132418+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:41:26.244820723+08:00","level":"INFO","msg":"stream: created new stream","id":"tfqi7r4k"} +{"time":"2025-09-16T16:41:26.244907278+08:00","level":"INFO","msg":"stream: started","id":"tfqi7r4k"} +{"time":"2025-09-16T16:41:26.244939252+08:00","level":"INFO","msg":"sender: started","stream_id":"tfqi7r4k"} +{"time":"2025-09-16T16:41:26.244934161+08:00","level":"INFO","msg":"writer: started","stream_id":"tfqi7r4k"} +{"time":"2025-09-16T16:41:26.244951691+08:00","level":"INFO","msg":"handler: started","stream_id":"tfqi7r4k"} +{"time":"2025-09-16T16:42:49.838878412+08:00","level":"INFO","msg":"stream: closing","id":"tfqi7r4k"} +{"time":"2025-09-16T16:42:51.179582102+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T16:42:51.435287371+08:00","level":"INFO","msg":"handler: closed","stream_id":"tfqi7r4k"} +{"time":"2025-09-16T16:42:51.435982972+08:00","level":"INFO","msg":"sender: closed","stream_id":"tfqi7r4k"} +{"time":"2025-09-16T16:42:51.436015682+08:00","level":"INFO","msg":"stream: closed","id":"tfqi7r4k"} diff --git a/wandb/run-20250916_164125-tfqi7r4k/logs/debug.log b/wandb/run-20250916_164125-tfqi7r4k/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..85d0076140f9af59619e6e3be29d01333be7c9fe --- /dev/null +++ b/wandb/run-20250916_164125-tfqi7r4k/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 16:41:25,552 INFO MainThread:3474582 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:41:25,552 INFO MainThread:3474582 [wandb_setup.py:_flush():81] Configure stats pid to 3474582 +2025-09-16 16:41:25,552 INFO MainThread:3474582 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:41:25,552 INFO MainThread:3474582 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:41:25,552 INFO MainThread:3474582 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:41:25,552 INFO MainThread:3474582 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_164125-tfqi7r4k/logs/debug.log +2025-09-16 16:41:25,553 INFO MainThread:3474582 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_164125-tfqi7r4k/logs/debug-internal.log +2025-09-16 16:41:25,553 INFO MainThread:3474582 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:41:25,553 INFO MainThread:3474582 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:41:25,553 INFO MainThread:3474582 [wandb_init.py:init():854] starting backend +2025-09-16 16:41:25,772 INFO MainThread:3474582 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:41:25,776 INFO MainThread:3474582 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:41:25,779 INFO MainThread:3474582 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:41:25,790 INFO MainThread:3474582 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:41:26,553 INFO MainThread:3474582 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:41:26,673 INFO MainThread:3474582 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:41:26,674 INFO MainThread:3474582 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:41:26,674 INFO MainThread:3474582 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:41:26,674 INFO MainThread:3474582 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:41:26,677 INFO MainThread:3474582 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:42:34,808 INFO MainThread:3474582 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 16:42:49,839 INFO wandb-AsyncioManager-main:3474582 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 16:42:49,839 INFO wandb-AsyncioManager-main:3474582 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_164125-tfqi7r4k/run-tfqi7r4k.wandb b/wandb/run-20250916_164125-tfqi7r4k/run-tfqi7r4k.wandb new file mode 100644 index 0000000000000000000000000000000000000000..61ce673d23c2013462efd4b109e93ff95f46a68b Binary files /dev/null and b/wandb/run-20250916_164125-tfqi7r4k/run-tfqi7r4k.wandb differ diff --git a/wandb/run-20250916_165221-qk4p7kad/files/output.log b/wandb/run-20250916_165221-qk4p7kad/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..9f4ab2fcfbc19523cef3907b45bbc16ab03ff22e --- /dev/null +++ b/wandb/run-20250916_165221-qk4p7kad/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:16:11, 0.04it/s, v_num=7kad] diff --git a/wandb/run-20250916_165221-qk4p7kad/files/requirements.txt b/wandb/run-20250916_165221-qk4p7kad/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_165221-qk4p7kad/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_165221-qk4p7kad/files/wandb-metadata.json b/wandb/run-20250916_165221-qk4p7kad/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..66678b71f83e5abf2b43f554a9f9b23cde56b982 --- /dev/null +++ b/wandb/run-20250916_165221-qk4p7kad/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T08:52:21.751062Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577058074624" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "nz35b7wanjpe8xvwuovuh5z4kxvwnjcx" +} \ No newline at end of file diff --git a/wandb/run-20250916_165221-qk4p7kad/logs/debug-core.log b/wandb/run-20250916_165221-qk4p7kad/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..8b75c83f326b3b1adddd91d75cb28318e8b87c39 --- /dev/null +++ b/wandb/run-20250916_165221-qk4p7kad/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T16:52:21.787315092+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpgd7dxh58/port-3500524.txt","pid":3500524,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T16:52:21.787539799+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T16:52:21.788208072+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3500524} +{"time":"2025-09-16T16:52:21.78814178+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3500524-3501549-2744912821/socket","Net":"unix"}} +{"time":"2025-09-16T16:52:21.9753774+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T16:52:21.985761178+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"qk4p7kad","id":"1(@)"} +{"time":"2025-09-16T16:52:22.454489406+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"qk4p7kad","id":"1(@)"} +{"time":"2025-09-16T16:55:09.13914636+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_165221-qk4p7kad/logs/debug-internal.log b/wandb/run-20250916_165221-qk4p7kad/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8b5df65b3d4c86fedaac944cd8ce0cbbbbcd4223 --- /dev/null +++ b/wandb/run-20250916_165221-qk4p7kad/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T16:52:21.986045973+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T16:52:22.006313116+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T16:52:22.454369157+08:00","level":"INFO","msg":"stream: created new stream","id":"qk4p7kad"} +{"time":"2025-09-16T16:52:22.454474873+08:00","level":"INFO","msg":"stream: started","id":"qk4p7kad"} +{"time":"2025-09-16T16:52:22.454523453+08:00","level":"INFO","msg":"writer: started","stream_id":"qk4p7kad"} +{"time":"2025-09-16T16:52:22.454578251+08:00","level":"INFO","msg":"sender: started","stream_id":"qk4p7kad"} +{"time":"2025-09-16T16:52:22.454547419+08:00","level":"INFO","msg":"handler: started","stream_id":"qk4p7kad"} diff --git a/wandb/run-20250916_165221-qk4p7kad/logs/debug.log b/wandb/run-20250916_165221-qk4p7kad/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..49cc16f3d3990ceebb61adbab76f4380fee5e1b6 --- /dev/null +++ b/wandb/run-20250916_165221-qk4p7kad/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 16:52:21,754 INFO MainThread:3500524 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_setup.py:_flush():81] Configure stats pid to 3500524 +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_165221-qk4p7kad/logs/debug.log +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_165221-qk4p7kad/logs/debug-internal.log +2025-09-16 16:52:21,755 INFO MainThread:3500524 [wandb_init.py:init():813] calling init triggers +2025-09-16 16:52:21,756 INFO MainThread:3500524 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 16:52:21,756 INFO MainThread:3500524 [wandb_init.py:init():854] starting backend +2025-09-16 16:52:21,975 INFO MainThread:3500524 [wandb_init.py:init():857] sending inform_init request +2025-09-16 16:52:21,980 INFO MainThread:3500524 [wandb_init.py:init():865] backend started and connected +2025-09-16 16:52:21,983 INFO MainThread:3500524 [wandb_init.py:init():936] updated telemetry +2025-09-16 16:52:21,993 INFO MainThread:3500524 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 16:52:22,885 INFO MainThread:3500524 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 16:52:23,066 INFO MainThread:3500524 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 16:52:23,066 INFO MainThread:3500524 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 16:52:23,066 INFO MainThread:3500524 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 16:52:23,066 INFO MainThread:3500524 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 16:52:23,069 INFO MainThread:3500524 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 16:53:31,837 INFO MainThread:3500524 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_165221-qk4p7kad/run-qk4p7kad.wandb b/wandb/run-20250916_165221-qk4p7kad/run-qk4p7kad.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_170107-ajwzo482/files/config.yaml b/wandb/run-20250916_170107-ajwzo482/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..61f19dfbd0c2cb24d3f367b598b9a2ae305a49a1 --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + m7qaebw7axnjktqn2gov0hxtvvs82n99: + args: + - fit + - --data=WordTaboo + - --data.batch_size=1 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577058766848" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T09:01:07.732210Z" + writerId: m7qaebw7axnjktqn2gov0hxtvvs82n99 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_170107-ajwzo482/files/output.log b/wandb/run-20250916_170107-ajwzo482/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e316c8ca8fc2e65f58f0460f209bce0126bc1930 --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/files/output.log @@ -0,0 +1,12 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 13/3305 [03:56<16:36:08, 0.06it/s, v_num=o482] +[rank: 0] Received SIGTERM: 15 + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_170107-ajwzo482/files/requirements.txt b/wandb/run-20250916_170107-ajwzo482/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_170107-ajwzo482/files/wandb-metadata.json b/wandb/run-20250916_170107-ajwzo482/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ff73bdb3533c9964fb72d7491059f2a9a2bbd881 --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:01:07.732210Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=1", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577058766848" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "m7qaebw7axnjktqn2gov0hxtvvs82n99" +} \ No newline at end of file diff --git a/wandb/run-20250916_170107-ajwzo482/files/wandb-summary.json b/wandb/run-20250916_170107-ajwzo482/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e1ae3765d42e07570d3107d18ffc33e950e07e62 --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":323,"_wandb":{"runtime":323}} \ No newline at end of file diff --git a/wandb/run-20250916_170107-ajwzo482/logs/debug-core.log b/wandb/run-20250916_170107-ajwzo482/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..51b70cd07ec6f30fed6b0cb291498ef81121c84a --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T17:01:07.767482681+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpp01996yx/port-3521737.txt","pid":3521737,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:01:07.767679339+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:01:07.768270226+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3521737} +{"time":"2025-09-16T17:01:07.768251942+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3521737-3522480-2094915690/socket","Net":"unix"}} +{"time":"2025-09-16T17:01:07.956647559+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:01:07.964359307+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ajwzo482","id":"1(@)"} +{"time":"2025-09-16T17:01:08.427351248+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ajwzo482","id":"1(@)"} +{"time":"2025-09-16T17:06:32.18194217+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T17:06:32.182120048+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T17:06:32.182111911+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T17:06:32.182262794+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T17:06:32.182317554+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3521737-3522480-2094915690/socket","Net":"unix"}} +{"time":"2025-09-16T17:06:33.776660091+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T17:06:33.776704831+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T17:06:33.776722731+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_170107-ajwzo482/logs/debug-internal.log b/wandb/run-20250916_170107-ajwzo482/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..959e5b339b66021ea8af13e271553cc45970fde2 --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T17:01:07.964511141+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:01:07.986088636+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:01:08.42721778+08:00","level":"INFO","msg":"stream: created new stream","id":"ajwzo482"} +{"time":"2025-09-16T17:01:08.427335387+08:00","level":"INFO","msg":"stream: started","id":"ajwzo482"} +{"time":"2025-09-16T17:01:08.42735399+08:00","level":"INFO","msg":"handler: started","stream_id":"ajwzo482"} +{"time":"2025-09-16T17:01:08.427409498+08:00","level":"INFO","msg":"writer: started","stream_id":"ajwzo482"} +{"time":"2025-09-16T17:01:08.427396645+08:00","level":"INFO","msg":"sender: started","stream_id":"ajwzo482"} +{"time":"2025-09-16T17:06:32.182099473+08:00","level":"INFO","msg":"stream: closing","id":"ajwzo482"} +{"time":"2025-09-16T17:06:33.501873252+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T17:06:33.77529912+08:00","level":"INFO","msg":"handler: closed","stream_id":"ajwzo482"} +{"time":"2025-09-16T17:06:33.775947623+08:00","level":"INFO","msg":"sender: closed","stream_id":"ajwzo482"} +{"time":"2025-09-16T17:06:33.775986673+08:00","level":"INFO","msg":"stream: closed","id":"ajwzo482"} diff --git a/wandb/run-20250916_170107-ajwzo482/logs/debug.log b/wandb/run-20250916_170107-ajwzo482/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..647b629d5a8a8310a47c53e4588efcd2ffb8459b --- /dev/null +++ b/wandb/run-20250916_170107-ajwzo482/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 17:01:07,735 INFO MainThread:3521737 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_setup.py:_flush():81] Configure stats pid to 3521737 +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_170107-ajwzo482/logs/debug.log +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_170107-ajwzo482/logs/debug-internal.log +2025-09-16 17:01:07,736 INFO MainThread:3521737 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:01:07,737 INFO MainThread:3521737 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:01:07,737 INFO MainThread:3521737 [wandb_init.py:init():854] starting backend +2025-09-16 17:01:07,956 INFO MainThread:3521737 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:01:07,961 INFO MainThread:3521737 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:01:07,964 INFO MainThread:3521737 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:01:07,974 INFO MainThread:3521737 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:01:08,767 INFO MainThread:3521737 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:01:08,932 INFO MainThread:3521737 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:01:08,932 INFO MainThread:3521737 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:01:08,933 INFO MainThread:3521737 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:01:08,933 INFO MainThread:3521737 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:01:08,935 INFO MainThread:3521737 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:02:18,733 INFO MainThread:3521737 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 17:06:32,181 INFO wandb-AsyncioManager-main:3521737 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 17:06:32,182 INFO wandb-AsyncioManager-main:3521737 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_170107-ajwzo482/run-ajwzo482.wandb b/wandb/run-20250916_170107-ajwzo482/run-ajwzo482.wandb new file mode 100644 index 0000000000000000000000000000000000000000..bd343dfe55823fb04c5f1f2133da47f77fa165b2 Binary files /dev/null and b/wandb/run-20250916_170107-ajwzo482/run-ajwzo482.wandb differ diff --git a/wandb/run-20250916_170718-nq5ffpyd/files/output.log b/wandb/run-20250916_170718-nq5ffpyd/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..dc1d8fc0c13ab5d8814e23b36dd83841bc908ad3 --- /dev/null +++ b/wandb/run-20250916_170718-nq5ffpyd/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:10:29, 0.04it/s, v_num=fpyd] diff --git a/wandb/run-20250916_170718-nq5ffpyd/files/requirements.txt b/wandb/run-20250916_170718-nq5ffpyd/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_170718-nq5ffpyd/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_170718-nq5ffpyd/files/wandb-metadata.json b/wandb/run-20250916_170718-nq5ffpyd/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4f5beafabe301fcdee2898b70b6ffe1bf785c033 --- /dev/null +++ b/wandb/run-20250916_170718-nq5ffpyd/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:07:18.035914Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577059373056" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "i7kurbwc5ndyk29sve54lus25b08rqa0" +} \ No newline at end of file diff --git a/wandb/run-20250916_170718-nq5ffpyd/logs/debug-core.log b/wandb/run-20250916_170718-nq5ffpyd/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..04e1d74909d06e8ddcdb162b9f3d6ba14a83407f --- /dev/null +++ b/wandb/run-20250916_170718-nq5ffpyd/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T17:07:18.071189955+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpowmdus0m/port-3537064.txt","pid":3537064,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:07:18.071393388+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:07:18.071961575+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3537064} +{"time":"2025-09-16T17:07:18.071969843+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3537064-3537810-3425285849/socket","Net":"unix"}} +{"time":"2025-09-16T17:07:18.260316549+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:07:18.276900036+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"nq5ffpyd","id":"1(@)"} +{"time":"2025-09-16T17:07:18.773252176+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"nq5ffpyd","id":"1(@)"} +{"time":"2025-09-16T17:10:06.240854649+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_170718-nq5ffpyd/logs/debug-internal.log b/wandb/run-20250916_170718-nq5ffpyd/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b785042847a0158c172dc7ef8dcc5c0a16fe4ec2 --- /dev/null +++ b/wandb/run-20250916_170718-nq5ffpyd/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T17:07:18.277175258+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:07:18.314697686+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:07:18.773139863+08:00","level":"INFO","msg":"stream: created new stream","id":"nq5ffpyd"} +{"time":"2025-09-16T17:07:18.773234903+08:00","level":"INFO","msg":"stream: started","id":"nq5ffpyd"} +{"time":"2025-09-16T17:07:18.773353468+08:00","level":"INFO","msg":"writer: started","stream_id":"nq5ffpyd"} +{"time":"2025-09-16T17:07:18.773500357+08:00","level":"INFO","msg":"sender: started","stream_id":"nq5ffpyd"} +{"time":"2025-09-16T17:07:18.773516912+08:00","level":"INFO","msg":"handler: started","stream_id":"nq5ffpyd"} diff --git a/wandb/run-20250916_170718-nq5ffpyd/logs/debug.log b/wandb/run-20250916_170718-nq5ffpyd/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..72831e6f2a69928c3200b31c86ef86f94d59cccc --- /dev/null +++ b/wandb/run-20250916_170718-nq5ffpyd/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 17:07:18,039 INFO MainThread:3537064 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_setup.py:_flush():81] Configure stats pid to 3537064 +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_170718-nq5ffpyd/logs/debug.log +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_170718-nq5ffpyd/logs/debug-internal.log +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:07:18,040 INFO MainThread:3537064 [wandb_init.py:init():854] starting backend +2025-09-16 17:07:18,260 INFO MainThread:3537064 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:07:18,265 INFO MainThread:3537064 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:07:18,268 INFO MainThread:3537064 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:07:18,279 INFO MainThread:3537064 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:07:19,115 INFO MainThread:3537064 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:07:19,300 INFO MainThread:3537064 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:07:19,300 INFO MainThread:3537064 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:07:19,301 INFO MainThread:3537064 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:07:19,301 INFO MainThread:3537064 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:07:19,304 INFO MainThread:3537064 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:08:28,340 INFO MainThread:3537064 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_170718-nq5ffpyd/run-nq5ffpyd.wandb b/wandb/run-20250916_170718-nq5ffpyd/run-nq5ffpyd.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_171259-ni87jpnx/files/config.yaml b/wandb/run-20250916_171259-ni87jpnx/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6031d023b3c3a5cc39197c84862f7861b391f922 --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 5mmfqksenwoene0wnwy4plfg1xggtqdf: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577060044800" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T09:12:59.476459Z" + writerId: 5mmfqksenwoene0wnwy4plfg1xggtqdf + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_171259-ni87jpnx/files/output.log b/wandb/run-20250916_171259-ni87jpnx/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..a9feb2db1fe3a7a8e81eb93549ad728ddcebe543 --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/files/output.log @@ -0,0 +1,139 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 551, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 629, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 633, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 145, in get_logsum_prob + last_hidden = obs_outputs.last_hidden_state[:, -1, :] +AttributeError: 'CausalLMOutputWithPast' object has no attribute 'last_hidden_state' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 551, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 629, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 633, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 145, in get_logsum_prob +[rank0]: last_hidden = obs_outputs.last_hidden_state[:, -1, :] +[rank0]: AttributeError: 'CausalLMOutputWithPast' object has no attribute 'last_hidden_state' diff --git a/wandb/run-20250916_171259-ni87jpnx/files/requirements.txt b/wandb/run-20250916_171259-ni87jpnx/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_171259-ni87jpnx/files/wandb-metadata.json b/wandb/run-20250916_171259-ni87jpnx/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c2606d4323881ab7927f094616fb1f0af2ee9fe0 --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:12:59.476459Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577060044800" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "5mmfqksenwoene0wnwy4plfg1xggtqdf" +} \ No newline at end of file diff --git a/wandb/run-20250916_171259-ni87jpnx/files/wandb-summary.json b/wandb/run-20250916_171259-ni87jpnx/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..2d0639155343dcebf02dc833d4e9f998d2b980a2 --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":83,"_wandb":{"runtime":83}} \ No newline at end of file diff --git a/wandb/run-20250916_171259-ni87jpnx/logs/debug-core.log b/wandb/run-20250916_171259-ni87jpnx/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bfe647cffd28b312bf90b6104ccc8184b483aeef --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T17:12:59.511417927+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp90csftcq/port-3551642.txt","pid":3551642,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:12:59.511715192+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:12:59.513073089+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3551642} +{"time":"2025-09-16T17:12:59.513083591+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3551642-3552316-2812618625/socket","Net":"unix"}} +{"time":"2025-09-16T17:12:59.700366686+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:12:59.708431941+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ni87jpnx","id":"1(@)"} +{"time":"2025-09-16T17:13:00.187346637+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ni87jpnx","id":"1(@)"} +{"time":"2025-09-16T17:14:27.741597399+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T17:14:27.741723579+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T17:14:27.741701848+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T17:14:27.74198909+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T17:14:27.742053279+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3551642-3552316-2812618625/socket","Net":"unix"}} +{"time":"2025-09-16T17:14:28.660554922+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T17:14:28.660604033+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T17:14:28.660634257+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_171259-ni87jpnx/logs/debug-internal.log b/wandb/run-20250916_171259-ni87jpnx/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..53cb4ae13e3c43ad68cbedc72f050c16460ccada --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T17:12:59.708586734+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:12:59.729327726+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:13:00.187214263+08:00","level":"INFO","msg":"stream: created new stream","id":"ni87jpnx"} +{"time":"2025-09-16T17:13:00.187331184+08:00","level":"INFO","msg":"stream: started","id":"ni87jpnx"} +{"time":"2025-09-16T17:13:00.187363294+08:00","level":"INFO","msg":"writer: started","stream_id":"ni87jpnx"} +{"time":"2025-09-16T17:13:00.18750857+08:00","level":"INFO","msg":"sender: started","stream_id":"ni87jpnx"} +{"time":"2025-09-16T17:13:00.187534838+08:00","level":"INFO","msg":"handler: started","stream_id":"ni87jpnx"} +{"time":"2025-09-16T17:14:27.7416978+08:00","level":"INFO","msg":"stream: closing","id":"ni87jpnx"} +{"time":"2025-09-16T17:14:28.377131503+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T17:14:28.658995133+08:00","level":"INFO","msg":"handler: closed","stream_id":"ni87jpnx"} +{"time":"2025-09-16T17:14:28.659517105+08:00","level":"INFO","msg":"sender: closed","stream_id":"ni87jpnx"} +{"time":"2025-09-16T17:14:28.659589313+08:00","level":"INFO","msg":"stream: closed","id":"ni87jpnx"} diff --git a/wandb/run-20250916_171259-ni87jpnx/logs/debug.log b/wandb/run-20250916_171259-ni87jpnx/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..0d606449158a2198917532ac5f242b106b7f107e --- /dev/null +++ b/wandb/run-20250916_171259-ni87jpnx/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 17:12:59,480 INFO MainThread:3551642 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:12:59,480 INFO MainThread:3551642 [wandb_setup.py:_flush():81] Configure stats pid to 3551642 +2025-09-16 17:12:59,480 INFO MainThread:3551642 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:12:59,480 INFO MainThread:3551642 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:12:59,481 INFO MainThread:3551642 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:12:59,481 INFO MainThread:3551642 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_171259-ni87jpnx/logs/debug.log +2025-09-16 17:12:59,481 INFO MainThread:3551642 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_171259-ni87jpnx/logs/debug-internal.log +2025-09-16 17:12:59,481 INFO MainThread:3551642 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:12:59,481 INFO MainThread:3551642 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:12:59,481 INFO MainThread:3551642 [wandb_init.py:init():854] starting backend +2025-09-16 17:12:59,700 INFO MainThread:3551642 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:12:59,705 INFO MainThread:3551642 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:12:59,708 INFO MainThread:3551642 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:12:59,719 INFO MainThread:3551642 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:13:04,586 INFO MainThread:3551642 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:13:04,749 INFO MainThread:3551642 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:13:04,749 INFO MainThread:3551642 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:13:04,753 INFO MainThread:3551642 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:13:04,753 INFO MainThread:3551642 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:13:04,756 INFO MainThread:3551642 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:14:14,087 INFO MainThread:3551642 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 17:14:27,743 INFO wandb-AsyncioManager-main:3551642 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 17:14:27,744 INFO wandb-AsyncioManager-main:3551642 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_171259-ni87jpnx/run-ni87jpnx.wandb b/wandb/run-20250916_171259-ni87jpnx/run-ni87jpnx.wandb new file mode 100644 index 0000000000000000000000000000000000000000..a164bf091bfb408a12a9027533663dce1750e091 Binary files /dev/null and b/wandb/run-20250916_171259-ni87jpnx/run-ni87jpnx.wandb differ diff --git a/wandb/run-20250916_171924-pnx0fzab/files/config.yaml b/wandb/run-20250916_171924-pnx0fzab/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..08737008919752a564c1f4101bfc88515d1d704b --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 7th4e48zp66fs41xvjt3j62ekkirzfsv: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577060499456" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T09:19:24.417847Z" + writerId: 7th4e48zp66fs41xvjt3j62ekkirzfsv + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_171924-pnx0fzab/files/output.log b/wandb/run-20250916_171924-pnx0fzab/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..669c499a5932aae2b173ea95b75319fc0f932fc4 --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/files/output.log @@ -0,0 +1,249 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 550, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 628, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 632, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 164, in get_logsum_prob + token_output = self.model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward + return self.base_model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward + return self.model.forward(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 940, in wrapper + output = func(self, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward + outputs: BaseModelOutputWithPast = self.model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 1064, in wrapper + outputs = func(self, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 410, in forward + hidden_states = decoder_layer( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ + return super().__call__(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 260, in forward + hidden_states, _ = self.self_attn( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 216, in forward + attn_output, attn_weights = attention_interface( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 83, in sdpa_attention_forward + attn_output = torch.nn.functional.scaled_dot_product_attention( +RuntimeError: Expected query, key, and value to have the same dtype, but got query.dtype: c10::BFloat16 key.dtype: float and value.dtype: float instead. +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 550, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 628, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 632, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 164, in get_logsum_prob +[rank0]: token_output = self.model( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward +[rank0]: return self.base_model( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward +[rank0]: return self.model.forward(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 940, in wrapper +[rank0]: output = func(self, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward +[rank0]: outputs: BaseModelOutputWithPast = self.model( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 1064, in wrapper +[rank0]: outputs = func(self, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 410, in forward +[rank0]: hidden_states = decoder_layer( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ +[rank0]: return super().__call__(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func +[rank0]: return func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 260, in forward +[rank0]: hidden_states, _ = self.self_attn( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func +[rank0]: return func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 216, in forward +[rank0]: attn_output, attn_weights = attention_interface( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 83, in sdpa_attention_forward +[rank0]: attn_output = torch.nn.functional.scaled_dot_product_attention( +[rank0]: RuntimeError: Expected query, key, and value to have the same dtype, but got query.dtype: c10::BFloat16 key.dtype: float and value.dtype: float instead. diff --git a/wandb/run-20250916_171924-pnx0fzab/files/requirements.txt b/wandb/run-20250916_171924-pnx0fzab/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_171924-pnx0fzab/files/wandb-metadata.json b/wandb/run-20250916_171924-pnx0fzab/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c750ed8ba4bb13e2ece87024ec10e1f3ea3cc87f --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:19:24.417847Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577060499456" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "7th4e48zp66fs41xvjt3j62ekkirzfsv" +} \ No newline at end of file diff --git a/wandb/run-20250916_171924-pnx0fzab/files/wandb-summary.json b/wandb/run-20250916_171924-pnx0fzab/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..c90edda74f6e5a2a5eb008d0c2f1dd0de08aa4f7 --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":82,"_wandb":{"runtime":82}} \ No newline at end of file diff --git a/wandb/run-20250916_171924-pnx0fzab/logs/debug-core.log b/wandb/run-20250916_171924-pnx0fzab/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..cfa1d9e78853703a03629722f510bb227cc28412 --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T17:19:24.452942341+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpo2hm0f7b/port-3567027.txt","pid":3567027,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:19:24.453259783+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:19:24.454117946+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3567027} +{"time":"2025-09-16T17:19:24.454110085+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3567027-3568274-2913859057/socket","Net":"unix"}} +{"time":"2025-09-16T17:19:24.640546027+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:19:24.648762726+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"pnx0fzab","id":"1(@)"} +{"time":"2025-09-16T17:19:25.125709434+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"pnx0fzab","id":"1(@)"} +{"time":"2025-09-16T17:20:48.433360419+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T17:20:48.433453975+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T17:20:48.433441516+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T17:20:48.433569017+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T17:20:48.433614003+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3567027-3568274-2913859057/socket","Net":"unix"}} +{"time":"2025-09-16T17:20:49.60267278+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T17:20:49.602733087+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T17:20:49.602765775+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_171924-pnx0fzab/logs/debug-internal.log b/wandb/run-20250916_171924-pnx0fzab/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..1db626a79e665a1f5c61bd4d339a0541f411fd14 --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T17:19:24.64894533+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:19:24.674031082+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:19:25.125562968+08:00","level":"INFO","msg":"stream: created new stream","id":"pnx0fzab"} +{"time":"2025-09-16T17:19:25.125693469+08:00","level":"INFO","msg":"stream: started","id":"pnx0fzab"} +{"time":"2025-09-16T17:19:25.125761484+08:00","level":"INFO","msg":"writer: started","stream_id":"pnx0fzab"} +{"time":"2025-09-16T17:19:25.125906064+08:00","level":"INFO","msg":"handler: started","stream_id":"pnx0fzab"} +{"time":"2025-09-16T17:19:25.126004962+08:00","level":"INFO","msg":"sender: started","stream_id":"pnx0fzab"} +{"time":"2025-09-16T17:20:48.433447427+08:00","level":"INFO","msg":"stream: closing","id":"pnx0fzab"} +{"time":"2025-09-16T17:20:49.241596026+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T17:20:49.601024461+08:00","level":"INFO","msg":"handler: closed","stream_id":"pnx0fzab"} +{"time":"2025-09-16T17:20:49.601519752+08:00","level":"INFO","msg":"sender: closed","stream_id":"pnx0fzab"} +{"time":"2025-09-16T17:20:49.601571456+08:00","level":"INFO","msg":"stream: closed","id":"pnx0fzab"} diff --git a/wandb/run-20250916_171924-pnx0fzab/logs/debug.log b/wandb/run-20250916_171924-pnx0fzab/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7329b06a8b9fba3663292e9bba20acad5df79209 --- /dev/null +++ b/wandb/run-20250916_171924-pnx0fzab/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 17:19:24,421 INFO MainThread:3567027 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:19:24,421 INFO MainThread:3567027 [wandb_setup.py:_flush():81] Configure stats pid to 3567027 +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_171924-pnx0fzab/logs/debug.log +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_171924-pnx0fzab/logs/debug-internal.log +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:19:24,422 INFO MainThread:3567027 [wandb_init.py:init():854] starting backend +2025-09-16 17:19:24,640 INFO MainThread:3567027 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:19:24,645 INFO MainThread:3567027 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:19:24,647 INFO MainThread:3567027 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:19:24,658 INFO MainThread:3567027 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:19:25,569 INFO MainThread:3567027 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:19:25,684 INFO MainThread:3567027 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:19:25,686 INFO MainThread:3567027 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:19:25,686 INFO MainThread:3567027 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:19:25,686 INFO MainThread:3567027 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:19:25,687 INFO MainThread:3567027 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:20:34,115 INFO MainThread:3567027 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 17:20:48,433 INFO wandb-AsyncioManager-main:3567027 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 17:20:48,433 INFO wandb-AsyncioManager-main:3567027 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_171924-pnx0fzab/run-pnx0fzab.wandb b/wandb/run-20250916_171924-pnx0fzab/run-pnx0fzab.wandb new file mode 100644 index 0000000000000000000000000000000000000000..d0f86b9fad33606b32d786a3042b800f74135cbe Binary files /dev/null and b/wandb/run-20250916_171924-pnx0fzab/run-pnx0fzab.wandb differ diff --git a/wandb/run-20250916_172412-fvdcryhs/files/config.yaml b/wandb/run-20250916_172412-fvdcryhs/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6a0640c4a2faf4b07ceb5a5d98aa3646b853b292 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + 8uflaiweyn59vtit25ys154i5drbek44: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577061048320" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T09:24:12.073783Z" + writerId: 8uflaiweyn59vtit25ys154i5drbek44 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_172412-fvdcryhs/files/output.log b/wandb/run-20250916_172412-fvdcryhs/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..669c499a5932aae2b173ea95b75319fc0f932fc4 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/files/output.log @@ -0,0 +1,249 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 550, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 628, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 632, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 164, in get_logsum_prob + token_output = self.model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward + return self.base_model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward + return self.model.forward(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 940, in wrapper + output = func(self, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward + outputs: BaseModelOutputWithPast = self.model( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 1064, in wrapper + outputs = func(self, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 410, in forward + hidden_states = decoder_layer( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ + return super().__call__(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 260, in forward + hidden_states, _ = self.self_attn( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func + return func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 216, in forward + attn_output, attn_weights = attention_interface( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 83, in sdpa_attention_forward + attn_output = torch.nn.functional.scaled_dot_product_attention( +RuntimeError: Expected query, key, and value to have the same dtype, but got query.dtype: c10::BFloat16 key.dtype: float and value.dtype: float instead. +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 550, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 628, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 632, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 164, in get_logsum_prob +[rank0]: token_output = self.model( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward +[rank0]: return self.base_model( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward +[rank0]: return self.model.forward(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 940, in wrapper +[rank0]: output = func(self, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 480, in forward +[rank0]: outputs: BaseModelOutputWithPast = self.model( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/generic.py", line 1064, in wrapper +[rank0]: outputs = func(self, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 410, in forward +[rank0]: hidden_states = decoder_layer( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/modeling_layers.py", line 94, in __call__ +[rank0]: return super().__call__(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func +[rank0]: return func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 260, in forward +[rank0]: hidden_states, _ = self.self_attn( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/utils/deprecation.py", line 172, in wrapped_func +[rank0]: return func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/models/qwen3/modeling_qwen3.py", line 216, in forward +[rank0]: attn_output, attn_weights = attention_interface( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/transformers/integrations/sdpa_attention.py", line 83, in sdpa_attention_forward +[rank0]: attn_output = torch.nn.functional.scaled_dot_product_attention( +[rank0]: RuntimeError: Expected query, key, and value to have the same dtype, but got query.dtype: c10::BFloat16 key.dtype: float and value.dtype: float instead. diff --git a/wandb/run-20250916_172412-fvdcryhs/files/requirements.txt b/wandb/run-20250916_172412-fvdcryhs/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_172412-fvdcryhs/files/wandb-metadata.json b/wandb/run-20250916_172412-fvdcryhs/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..396a9ecffaf9ee4bb1a23b5f2da322a4cac907f3 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:24:12.073783Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577061048320" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "8uflaiweyn59vtit25ys154i5drbek44" +} \ No newline at end of file diff --git a/wandb/run-20250916_172412-fvdcryhs/files/wandb-summary.json b/wandb/run-20250916_172412-fvdcryhs/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..dcdfe068b7fce1b91b105a1cf2d7aa757e7d7661 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":82},"_runtime":82} \ No newline at end of file diff --git a/wandb/run-20250916_172412-fvdcryhs/logs/debug-core.log b/wandb/run-20250916_172412-fvdcryhs/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..c9d4327b34185c86eee7fc32cbf142c3ab22f353 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T17:24:12.108472586+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpiwkvjypo/port-3579199.txt","pid":3579199,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:24:12.108748244+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:24:12.109373061+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3579199} +{"time":"2025-09-16T17:24:12.109343877+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3579199-3579934-761459508/socket","Net":"unix"}} +{"time":"2025-09-16T17:24:12.296354063+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:24:12.303469567+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"fvdcryhs","id":"1(@)"} +{"time":"2025-09-16T17:24:12.753822562+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"fvdcryhs","id":"1(@)"} +{"time":"2025-09-16T17:25:36.017450907+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T17:25:36.017495261+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T17:25:36.017557858+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T17:25:36.017556204+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T17:25:36.017846903+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3579199-3579934-761459508/socket","Net":"unix"}} +{"time":"2025-09-16T17:25:37.631599292+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T17:25:37.631639025+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T17:25:37.631658303+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_172412-fvdcryhs/logs/debug-internal.log b/wandb/run-20250916_172412-fvdcryhs/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..40437d2b556602c249e7dc7b800b46bbd41374b1 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T17:24:12.303605978+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:24:12.324332449+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:24:12.753711402+08:00","level":"INFO","msg":"stream: created new stream","id":"fvdcryhs"} +{"time":"2025-09-16T17:24:12.753807512+08:00","level":"INFO","msg":"stream: started","id":"fvdcryhs"} +{"time":"2025-09-16T17:24:12.753845486+08:00","level":"INFO","msg":"writer: started","stream_id":"fvdcryhs"} +{"time":"2025-09-16T17:24:12.753965297+08:00","level":"INFO","msg":"handler: started","stream_id":"fvdcryhs"} +{"time":"2025-09-16T17:24:12.754000467+08:00","level":"INFO","msg":"sender: started","stream_id":"fvdcryhs"} +{"time":"2025-09-16T17:25:36.017505833+08:00","level":"INFO","msg":"stream: closing","id":"fvdcryhs"} +{"time":"2025-09-16T17:25:37.360967536+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T17:25:37.629669017+08:00","level":"INFO","msg":"handler: closed","stream_id":"fvdcryhs"} +{"time":"2025-09-16T17:25:37.630335245+08:00","level":"INFO","msg":"sender: closed","stream_id":"fvdcryhs"} +{"time":"2025-09-16T17:25:37.630377667+08:00","level":"INFO","msg":"stream: closed","id":"fvdcryhs"} diff --git a/wandb/run-20250916_172412-fvdcryhs/logs/debug.log b/wandb/run-20250916_172412-fvdcryhs/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..f4a5972a6667ea5e18aa512b96f5a402a2606388 --- /dev/null +++ b/wandb/run-20250916_172412-fvdcryhs/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 17:24:12,077 INFO MainThread:3579199 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:24:12,077 INFO MainThread:3579199 [wandb_setup.py:_flush():81] Configure stats pid to 3579199 +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_172412-fvdcryhs/logs/debug.log +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_172412-fvdcryhs/logs/debug-internal.log +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:24:12,078 INFO MainThread:3579199 [wandb_init.py:init():854] starting backend +2025-09-16 17:24:12,296 INFO MainThread:3579199 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:24:12,301 INFO MainThread:3579199 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:24:12,304 INFO MainThread:3579199 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:24:12,314 INFO MainThread:3579199 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:24:13,050 INFO MainThread:3579199 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:24:13,171 INFO MainThread:3579199 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:24:13,171 INFO MainThread:3579199 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:24:13,171 INFO MainThread:3579199 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:24:13,171 INFO MainThread:3579199 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:24:13,173 INFO MainThread:3579199 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:25:21,938 INFO MainThread:3579199 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 17:25:36,018 INFO wandb-AsyncioManager-main:3579199 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 17:25:36,018 INFO wandb-AsyncioManager-main:3579199 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_172412-fvdcryhs/run-fvdcryhs.wandb b/wandb/run-20250916_172412-fvdcryhs/run-fvdcryhs.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e1dc08d7126310f1c2cf32224e81e25830ead002 Binary files /dev/null and b/wandb/run-20250916_172412-fvdcryhs/run-fvdcryhs.wandb differ diff --git a/wandb/run-20250916_173043-nzsopkhe/files/config.yaml b/wandb/run-20250916_173043-nzsopkhe/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..fe219dc0f81bc2dbe3f838a29f2ff6fee09fa45e --- /dev/null +++ b/wandb/run-20250916_173043-nzsopkhe/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + azw19xjls19y4f54kztr9bg8qj35fujy: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577061572608" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T09:30:43.955116Z" + writerId: azw19xjls19y4f54kztr9bg8qj35fujy + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_173043-nzsopkhe/files/output.log b/wandb/run-20250916_173043-nzsopkhe/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c0af0443a51cb1a0d2c88e2f3b6f643fe5369f44 --- /dev/null +++ b/wandb/run-20250916_173043-nzsopkhe/files/output.log @@ -0,0 +1,11 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 532, in training_step + self.manual_backward(actor_loss) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward + self.trainer.strategy.backward(loss, None, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward + self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward + deepspeed_engine.backward(tensor, *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward + self._do_optimizer_backward(loss, retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward + self.optimizer.backward(loss, retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward + self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward + scaled_loss.backward(retain_graph=retain_graph) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward + torch.autograd.backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward + _engine_run_backward( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward + return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 532, in training_step +[rank0]: self.manual_backward(actor_loss) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/core/module.py", line 1116, in manual_backward +[rank0]: self.trainer.strategy.backward(loss, None, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 213, in backward +[rank0]: self.precision_plugin.backward(closure_loss, self.lightning_module, optimizer, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/plugins/precision/deepspeed.py", line 117, in backward +[rank0]: deepspeed_engine.backward(tensor, *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2298, in backward +[rank0]: self._do_optimizer_backward(loss, retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2244, in _do_optimizer_backward +[rank0]: self.optimizer.backward(loss, retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/zero/stage3.py", line 2305, in backward +[rank0]: self.loss_scaler.backward(loss.float(), retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/fp16/loss_scaler.py", line 65, in backward +[rank0]: scaled_loss.backward(retain_graph=retain_graph) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/_tensor.py", line 647, in backward +[rank0]: torch.autograd.backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/__init__.py", line 354, in backward +[rank0]: _engine_run_backward( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/autograd/graph.py", line 829, in _engine_run_backward +[rank0]: return Variable._execution_engine.run_backward( # Calls into the C++ engine to run the backward pass +[rank0]: RuntimeError: element 0 of tensors does not require grad and does not have a grad_fn diff --git a/wandb/run-20250916_173449-mxl5eh3s/files/requirements.txt b/wandb/run-20250916_173449-mxl5eh3s/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_173449-mxl5eh3s/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_173449-mxl5eh3s/files/wandb-metadata.json b/wandb/run-20250916_173449-mxl5eh3s/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7ed9f4e6f7bb98976c34d61b92e799d3df6662c5 --- /dev/null +++ b/wandb/run-20250916_173449-mxl5eh3s/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:34:49.412750Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577062191104" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "gbka7fpj2jpnu828967vya7p4xcgcqhx" +} \ No newline at end of file diff --git a/wandb/run-20250916_173449-mxl5eh3s/files/wandb-summary.json b/wandb/run-20250916_173449-mxl5eh3s/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..90a011666e145893a12ccd6473aa458e58c3daa3 --- /dev/null +++ b/wandb/run-20250916_173449-mxl5eh3s/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":90},"_runtime":90} \ No newline at end of file diff --git a/wandb/run-20250916_173449-mxl5eh3s/logs/debug-core.log b/wandb/run-20250916_173449-mxl5eh3s/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..5c7c9e9d08be01b7159281e3147e8b143042d429 --- /dev/null +++ b/wandb/run-20250916_173449-mxl5eh3s/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T17:34:49.449403908+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp6j7t40ou/port-3606704.txt","pid":3606704,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:34:49.449584484+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:34:49.450137187+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3606704} +{"time":"2025-09-16T17:34:49.450130155+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3606704-3607429-520119664/socket","Net":"unix"}} +{"time":"2025-09-16T17:34:49.637309227+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:34:49.64614422+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"mxl5eh3s","id":"1(@)"} +{"time":"2025-09-16T17:34:50.147453161+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"mxl5eh3s","id":"1(@)"} +{"time":"2025-09-16T17:36:21.325463101+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T17:36:21.325537615+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T17:36:21.325525429+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T17:36:21.325701878+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T17:36:21.325714768+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3606704-3607429-520119664/socket","Net":"unix"}} +{"time":"2025-09-16T17:36:22.537869573+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T17:36:22.537913439+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T17:36:22.537932937+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_173449-mxl5eh3s/logs/debug-internal.log b/wandb/run-20250916_173449-mxl5eh3s/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..c7e97afc60554ce65d9ff8e53ee569701ab5fb9a --- /dev/null +++ b/wandb/run-20250916_173449-mxl5eh3s/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T17:34:49.646353683+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:34:49.67451546+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:34:50.147377492+08:00","level":"INFO","msg":"stream: created new stream","id":"mxl5eh3s"} +{"time":"2025-09-16T17:34:50.147444144+08:00","level":"INFO","msg":"stream: started","id":"mxl5eh3s"} +{"time":"2025-09-16T17:34:50.147462121+08:00","level":"INFO","msg":"writer: started","stream_id":"mxl5eh3s"} +{"time":"2025-09-16T17:34:50.147483737+08:00","level":"INFO","msg":"sender: started","stream_id":"mxl5eh3s"} +{"time":"2025-09-16T17:34:50.14749733+08:00","level":"INFO","msg":"handler: started","stream_id":"mxl5eh3s"} +{"time":"2025-09-16T17:36:21.325538125+08:00","level":"INFO","msg":"stream: closing","id":"mxl5eh3s"} +{"time":"2025-09-16T17:36:22.113340186+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T17:36:22.536441763+08:00","level":"INFO","msg":"handler: closed","stream_id":"mxl5eh3s"} +{"time":"2025-09-16T17:36:22.537007015+08:00","level":"INFO","msg":"sender: closed","stream_id":"mxl5eh3s"} +{"time":"2025-09-16T17:36:22.537044898+08:00","level":"INFO","msg":"stream: closed","id":"mxl5eh3s"} diff --git a/wandb/run-20250916_173449-mxl5eh3s/logs/debug.log b/wandb/run-20250916_173449-mxl5eh3s/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..bfcf3811e106bf848bd90b5a69973402e99044d2 --- /dev/null +++ b/wandb/run-20250916_173449-mxl5eh3s/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 17:34:49,416 INFO MainThread:3606704 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:34:49,416 INFO MainThread:3606704 [wandb_setup.py:_flush():81] Configure stats pid to 3606704 +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_173449-mxl5eh3s/logs/debug.log +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_173449-mxl5eh3s/logs/debug-internal.log +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:34:49,417 INFO MainThread:3606704 [wandb_init.py:init():854] starting backend +2025-09-16 17:34:49,637 INFO MainThread:3606704 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:34:49,642 INFO MainThread:3606704 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:34:49,645 INFO MainThread:3606704 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:34:49,656 INFO MainThread:3606704 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:34:50,459 INFO MainThread:3606704 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:34:50,628 INFO MainThread:3606704 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:34:50,628 INFO MainThread:3606704 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:34:50,628 INFO MainThread:3606704 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:34:50,628 INFO MainThread:3606704 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:34:50,631 INFO MainThread:3606704 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:35:59,633 INFO MainThread:3606704 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 17:36:21,325 INFO wandb-AsyncioManager-main:3606704 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 17:36:21,326 INFO wandb-AsyncioManager-main:3606704 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_173449-mxl5eh3s/run-mxl5eh3s.wandb b/wandb/run-20250916_173449-mxl5eh3s/run-mxl5eh3s.wandb new file mode 100644 index 0000000000000000000000000000000000000000..4c536ced1504eab45ff539a4cccf366e50072fb5 Binary files /dev/null and b/wandb/run-20250916_173449-mxl5eh3s/run-mxl5eh3s.wandb differ diff --git a/wandb/run-20250916_173908-s3om1bej/files/output.log b/wandb/run-20250916_173908-s3om1bej/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..0699e06c830e1caa6b62e54faaa8590a921c9495 --- /dev/null +++ b/wandb/run-20250916_173908-s3om1bej/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:10:11, 0.04it/s, v_num=1bej] diff --git a/wandb/run-20250916_173908-s3om1bej/files/requirements.txt b/wandb/run-20250916_173908-s3om1bej/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_173908-s3om1bej/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_173908-s3om1bej/files/wandb-metadata.json b/wandb/run-20250916_173908-s3om1bej/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..fe90b1633483d648feb1f0ee0eb2525cf4a50633 --- /dev/null +++ b/wandb/run-20250916_173908-s3om1bej/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:39:08.077109Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577062563840" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "e7axnvgs83rj6154nrq97jy5ug4j707b" +} \ No newline at end of file diff --git a/wandb/run-20250916_173908-s3om1bej/logs/debug-core.log b/wandb/run-20250916_173908-s3om1bej/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..b9806eb2373aa887d5ab106bc803e8db88388637 --- /dev/null +++ b/wandb/run-20250916_173908-s3om1bej/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T17:39:08.112221141+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpef39c66y/port-3617976.txt","pid":3617976,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:39:08.11243282+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:39:08.112950381+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3617976} +{"time":"2025-09-16T17:39:08.112958132+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3617976-3618665-1774759303/socket","Net":"unix"}} +{"time":"2025-09-16T17:39:08.3013488+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:39:08.309669016+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"s3om1bej","id":"1(@)"} +{"time":"2025-09-16T17:39:08.785264918+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"s3om1bej","id":"1(@)"} +{"time":"2025-09-16T17:41:55.06689793+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_173908-s3om1bej/logs/debug-internal.log b/wandb/run-20250916_173908-s3om1bej/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..7045861f5056449cea82014750ff879c61dfad27 --- /dev/null +++ b/wandb/run-20250916_173908-s3om1bej/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T17:39:08.309840289+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:39:08.324936717+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:39:08.785156004+08:00","level":"INFO","msg":"stream: created new stream","id":"s3om1bej"} +{"time":"2025-09-16T17:39:08.785249418+08:00","level":"INFO","msg":"stream: started","id":"s3om1bej"} +{"time":"2025-09-16T17:39:08.785280015+08:00","level":"INFO","msg":"writer: started","stream_id":"s3om1bej"} +{"time":"2025-09-16T17:39:08.785289138+08:00","level":"INFO","msg":"sender: started","stream_id":"s3om1bej"} +{"time":"2025-09-16T17:39:08.785334588+08:00","level":"INFO","msg":"handler: started","stream_id":"s3om1bej"} diff --git a/wandb/run-20250916_173908-s3om1bej/logs/debug.log b/wandb/run-20250916_173908-s3om1bej/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a752f3871ad44f79be940b2173de7db935f93cc8 --- /dev/null +++ b/wandb/run-20250916_173908-s3om1bej/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_setup.py:_flush():81] Configure stats pid to 3617976 +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_173908-s3om1bej/logs/debug.log +2025-09-16 17:39:08,081 INFO MainThread:3617976 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_173908-s3om1bej/logs/debug-internal.log +2025-09-16 17:39:08,082 INFO MainThread:3617976 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:39:08,082 INFO MainThread:3617976 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:39:08,082 INFO MainThread:3617976 [wandb_init.py:init():854] starting backend +2025-09-16 17:39:08,301 INFO MainThread:3617976 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:39:08,306 INFO MainThread:3617976 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:39:08,309 INFO MainThread:3617976 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:39:08,319 INFO MainThread:3617976 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:39:09,232 INFO MainThread:3617976 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:39:09,400 INFO MainThread:3617976 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:39:09,401 INFO MainThread:3617976 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:39:09,402 INFO MainThread:3617976 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:39:09,402 INFO MainThread:3617976 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:39:09,404 INFO MainThread:3617976 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:40:18,399 INFO MainThread:3617976 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_173908-s3om1bej/run-s3om1bej.wandb b/wandb/run-20250916_173908-s3om1bej/run-s3om1bej.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_174436-ur1lf4al/files/config.yaml b/wandb/run-20250916_174436-ur1lf4al/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..325b24c118854441003dab5985142b11ff385d0b --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + otqrg9infp4f5ewo3ejh61ahvac2f3d7: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3577063186432" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T09:44:36.033003Z" + writerId: otqrg9infp4f5ewo3ejh61ahvac2f3d7 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_174436-ur1lf4al/files/output.log b/wandb/run-20250916_174436-ur1lf4al/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e29295b683cd14ef022f50c34ab1e95d581b3e22 --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/files/output.log @@ -0,0 +1,173 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5922440 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 520, in training_step + critic_loss, critic_log = self.critic_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 615, in + self.critic_loss = lambda batch: self.critic.iql_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 267, in iql_loss + q1, q2 = self.get_q(observation, action, detach_model=False) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 227, in get_q + return self.critic.get_q(observation, action, detach_model=detach_model) + File "/home/jiashuo/codes/OfflineArcher/ArcherCritic.py", line 58, in get_q + return self.critic1(lm_states), self.critic2(lm_states) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/container.py", line 244, in forward + input = module(input) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 125, in forward + return F.linear(input, self.weight, self.bias) +RuntimeError: mat1 and mat2 shapes cannot be multiplied (2x1536 and 736x368) +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 520, in training_step +[rank0]: critic_loss, critic_log = self.critic_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 615, in +[rank0]: self.critic_loss = lambda batch: self.critic.iql_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 267, in iql_loss +[rank0]: q1, q2 = self.get_q(observation, action, detach_model=False) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 227, in get_q +[rank0]: return self.critic.get_q(observation, action, detach_model=detach_model) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/ArcherCritic.py", line 58, in get_q +[rank0]: return self.critic1(lm_states), self.critic2(lm_states) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/container.py", line 244, in forward +[rank0]: input = module(input) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/linear.py", line 125, in forward +[rank0]: return F.linear(input, self.weight, self.bias) +[rank0]: RuntimeError: mat1 and mat2 shapes cannot be multiplied (2x1536 and 736x368) diff --git a/wandb/run-20250916_174436-ur1lf4al/files/requirements.txt b/wandb/run-20250916_174436-ur1lf4al/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_174436-ur1lf4al/files/wandb-metadata.json b/wandb/run-20250916_174436-ur1lf4al/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b3763a15ea99d1fdedd343e5cdf873d304279828 --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:44:36.033003Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577063186432" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "otqrg9infp4f5ewo3ejh61ahvac2f3d7" +} \ No newline at end of file diff --git a/wandb/run-20250916_174436-ur1lf4al/files/wandb-summary.json b/wandb/run-20250916_174436-ur1lf4al/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..9f39067485dc32618da84a815b6711331233c7e9 --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":75},"_runtime":75} \ No newline at end of file diff --git a/wandb/run-20250916_174436-ur1lf4al/logs/debug-core.log b/wandb/run-20250916_174436-ur1lf4al/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..c3d34173947a011b354a743a5be9094c905e7692 --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/logs/debug-core.log @@ -0,0 +1,16 @@ +{"time":"2025-09-16T17:44:36.059178958+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp89bo873l/port-3631356.txt","pid":3631356,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:44:36.059409893+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:44:36.060476134+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3631356} +{"time":"2025-09-16T17:44:36.060420289+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3631356-3632622-1010312862/socket","Net":"unix"}} +{"time":"2025-09-16T17:44:36.248135543+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:44:36.255452344+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ur1lf4al","id":"1(@)"} +{"time":"2025-09-16T17:44:36.747468418+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ur1lf4al","id":"1(@)"} +{"time":"2025-09-16T17:45:52.203798132+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T17:45:52.203857057+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T17:45:52.203878742+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T17:45:52.203911781+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T17:45:52.204179067+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3631356-3632622-1010312862/socket","Net":"unix"}} +{"time":"2025-09-16T17:45:52.442773039+08:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write unix /tmp/wandb-3631356-3632622-1010312862/socket->@: use of closed network connection","id":"1(@)"} +{"time":"2025-09-16T17:45:53.524486779+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T17:45:53.524553459+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T17:45:53.524592258+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_174436-ur1lf4al/logs/debug-internal.log b/wandb/run-20250916_174436-ur1lf4al/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2ecd1915bbb26729e4aeb620835cc31a1eb7cea0 --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T17:44:36.255743989+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:44:36.281765716+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:44:36.747364996+08:00","level":"INFO","msg":"stream: created new stream","id":"ur1lf4al"} +{"time":"2025-09-16T17:44:36.747452991+08:00","level":"INFO","msg":"stream: started","id":"ur1lf4al"} +{"time":"2025-09-16T17:44:36.747496967+08:00","level":"INFO","msg":"handler: started","stream_id":"ur1lf4al"} +{"time":"2025-09-16T17:44:36.747520331+08:00","level":"INFO","msg":"sender: started","stream_id":"ur1lf4al"} +{"time":"2025-09-16T17:44:36.747481381+08:00","level":"INFO","msg":"writer: started","stream_id":"ur1lf4al"} +{"time":"2025-09-16T17:45:52.203862064+08:00","level":"INFO","msg":"stream: closing","id":"ur1lf4al"} +{"time":"2025-09-16T17:45:53.273963884+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T17:45:53.52255415+08:00","level":"INFO","msg":"handler: closed","stream_id":"ur1lf4al"} +{"time":"2025-09-16T17:45:53.523201466+08:00","level":"INFO","msg":"sender: closed","stream_id":"ur1lf4al"} +{"time":"2025-09-16T17:45:53.523263087+08:00","level":"INFO","msg":"stream: closed","id":"ur1lf4al"} diff --git a/wandb/run-20250916_174436-ur1lf4al/logs/debug.log b/wandb/run-20250916_174436-ur1lf4al/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..cd5ba78f08bf387b7bdda3261007ffab284e732a --- /dev/null +++ b/wandb/run-20250916_174436-ur1lf4al/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 17:44:36,034 INFO MainThread:3631356 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_setup.py:_flush():81] Configure stats pid to 3631356 +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_174436-ur1lf4al/logs/debug.log +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_174436-ur1lf4al/logs/debug-internal.log +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:44:36,035 INFO MainThread:3631356 [wandb_init.py:init():854] starting backend +2025-09-16 17:44:36,248 INFO MainThread:3631356 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:44:36,250 INFO MainThread:3631356 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:44:36,252 INFO MainThread:3631356 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:44:36,259 INFO MainThread:3631356 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:44:37,052 INFO MainThread:3631356 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:44:37,180 INFO MainThread:3631356 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:44:37,180 INFO MainThread:3631356 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:44:37,181 INFO MainThread:3631356 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:44:37,181 INFO MainThread:3631356 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:44:37,183 INFO MainThread:3631356 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:45:50,053 INFO MainThread:3631356 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 17:45:52,203 INFO wandb-AsyncioManager-main:3631356 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 17:45:52,204 INFO wandb-AsyncioManager-main:3631356 [mailbox.py:close():137] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250916_174436-ur1lf4al/run-ur1lf4al.wandb b/wandb/run-20250916_174436-ur1lf4al/run-ur1lf4al.wandb new file mode 100644 index 0000000000000000000000000000000000000000..0708627a5200e258a00cc0e7ab4e9b448aeea93d Binary files /dev/null and b/wandb/run-20250916_174436-ur1lf4al/run-ur1lf4al.wandb differ diff --git a/wandb/run-20250916_174811-6asoihqp/files/output.log b/wandb/run-20250916_174811-6asoihqp/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..155ddd2cdeef9b2b51445ca80d7342d407425fbd --- /dev/null +++ b/wandb/run-20250916_174811-6asoihqp/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [0,1,2,3] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5922824 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:53<12:10:16, 0.04it/s, v_num=ihqp] diff --git a/wandb/run-20250916_174811-6asoihqp/files/requirements.txt b/wandb/run-20250916_174811-6asoihqp/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_174811-6asoihqp/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_174811-6asoihqp/files/wandb-metadata.json b/wandb/run-20250916_174811-6asoihqp/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..0b0627aa889a40e96db3a4226132016300e20f22 --- /dev/null +++ b/wandb/run-20250916_174811-6asoihqp/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T09:48:11.559941Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3577063493632" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "2df9enimpxhohihhvu3n7hkd7x5gesn7" +} \ No newline at end of file diff --git a/wandb/run-20250916_174811-6asoihqp/logs/debug-core.log b/wandb/run-20250916_174811-6asoihqp/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..583fb19f4a002d331ab76d6fb6416316ab1450f1 --- /dev/null +++ b/wandb/run-20250916_174811-6asoihqp/logs/debug-core.log @@ -0,0 +1,9 @@ +{"time":"2025-09-16T17:48:11.597633503+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpmtuw7zwj/port-3640677.txt","pid":3640677,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T17:48:11.597832789+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T17:48:11.598875601+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3640677} +{"time":"2025-09-16T17:48:11.598912278+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3640677-3641369-1193320635/socket","Net":"unix"}} +{"time":"2025-09-16T17:48:11.788064433+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T17:48:11.800990919+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"6asoihqp","id":"1(@)"} +{"time":"2025-09-16T17:48:12.293511685+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"6asoihqp","id":"1(@)"} +{"time":"2025-09-16T17:50:58.088452608+08:00","level":"ERROR","msg":"connection: fatal error reading connection","error":"read unix /tmp/wandb-3640677-3641369-1193320635/socket->@: read: connection reset by peer","id":"1(@)"} +{"time":"2025-09-16T17:50:58.267938045+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_174811-6asoihqp/logs/debug-internal.log b/wandb/run-20250916_174811-6asoihqp/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..3e9d9c8b72790f77b69502a47c045fdcbf04cc67 --- /dev/null +++ b/wandb/run-20250916_174811-6asoihqp/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T17:48:11.801145708+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T17:48:11.820305336+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T17:48:12.293395964+08:00","level":"INFO","msg":"stream: created new stream","id":"6asoihqp"} +{"time":"2025-09-16T17:48:12.293497135+08:00","level":"INFO","msg":"stream: started","id":"6asoihqp"} +{"time":"2025-09-16T17:48:12.293531376+08:00","level":"INFO","msg":"handler: started","stream_id":"6asoihqp"} +{"time":"2025-09-16T17:48:12.293526547+08:00","level":"INFO","msg":"writer: started","stream_id":"6asoihqp"} +{"time":"2025-09-16T17:48:12.293556302+08:00","level":"INFO","msg":"sender: started","stream_id":"6asoihqp"} diff --git a/wandb/run-20250916_174811-6asoihqp/logs/debug.log b/wandb/run-20250916_174811-6asoihqp/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..41b351ef9952f92ab78673d2e48c7c3b4a8c5a13 --- /dev/null +++ b/wandb/run-20250916_174811-6asoihqp/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 17:48:11,565 INFO MainThread:3640677 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_setup.py:_flush():81] Configure stats pid to 3640677 +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_174811-6asoihqp/logs/debug.log +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_174811-6asoihqp/logs/debug-internal.log +2025-09-16 17:48:11,566 INFO MainThread:3640677 [wandb_init.py:init():813] calling init triggers +2025-09-16 17:48:11,567 INFO MainThread:3640677 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 17:48:11,567 INFO MainThread:3640677 [wandb_init.py:init():854] starting backend +2025-09-16 17:48:11,788 INFO MainThread:3640677 [wandb_init.py:init():857] sending inform_init request +2025-09-16 17:48:11,793 INFO MainThread:3640677 [wandb_init.py:init():865] backend started and connected +2025-09-16 17:48:11,796 INFO MainThread:3640677 [wandb_init.py:init():936] updated telemetry +2025-09-16 17:48:11,808 INFO MainThread:3640677 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 17:48:12,596 INFO MainThread:3640677 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 17:48:12,790 INFO MainThread:3640677 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 17:48:12,790 INFO MainThread:3640677 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 17:48:12,790 INFO MainThread:3640677 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 17:48:12,790 INFO MainThread:3640677 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 17:48:12,794 INFO MainThread:3640677 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 17:49:21,409 INFO MainThread:3640677 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_174811-6asoihqp/run-6asoihqp.wandb b/wandb/run-20250916_174811-6asoihqp/run-6asoihqp.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_193511-phresaiu/files/output.log b/wandb/run-20250916_193511-phresaiu/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4809163d8507355f8198c6c2d823e2bbcb7641f6 --- /dev/null +++ b/wandb/run-20250916_193511-phresaiu/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5922824 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 561, in training_step + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 639, in + critic_checkpoint=critic_checkpoint, + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 643, in awr_loss + self.inv_temp = inv_temp + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 169, in get_logsum_prob + if act_ids[0, j] == self.tokenizer.pad_token_id: +AttributeError: 'tuple' object has no attribute 'size' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 561, in training_step +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 639, in +[rank0]: critic_checkpoint=critic_checkpoint, +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 643, in awr_loss +[rank0]: self.inv_temp = inv_temp +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 169, in get_logsum_prob +[rank0]: if act_ids[0, j] == self.tokenizer.pad_token_id: +[rank0]: AttributeError: 'tuple' object has no attribute 'size' diff --git a/wandb/run-20250916_193756-etkbn0o2/files/requirements.txt b/wandb/run-20250916_193756-etkbn0o2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_193756-etkbn0o2/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_193756-etkbn0o2/files/wandb-metadata.json b/wandb/run-20250916_193756-etkbn0o2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7a4f7fdc63c0c3927c28e2010c58f6c50cc86d32 --- /dev/null +++ b/wandb/run-20250916_193756-etkbn0o2/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T11:37:56.567732Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576975392768" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "2c9c7aw9sc2pitkpqu44m36gqgddtgpr" +} \ No newline at end of file diff --git a/wandb/run-20250916_193756-etkbn0o2/files/wandb-summary.json b/wandb/run-20250916_193756-etkbn0o2/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..dcdfe068b7fce1b91b105a1cf2d7aa757e7d7661 --- /dev/null +++ b/wandb/run-20250916_193756-etkbn0o2/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":82},"_runtime":82} \ No newline at end of file diff --git a/wandb/run-20250916_193756-etkbn0o2/logs/debug-core.log b/wandb/run-20250916_193756-etkbn0o2/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..ba4087f5270e24690520cab40498f68cd3f0bdfa --- /dev/null +++ b/wandb/run-20250916_193756-etkbn0o2/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T19:37:56.603839922+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpdoz21_8v/port-3884253.txt","pid":3884253,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T19:37:56.604019812+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T19:37:56.604631154+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3884253} +{"time":"2025-09-16T19:37:56.604607433+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3884253-3884884-4178420690/socket","Net":"unix"}} +{"time":"2025-09-16T19:37:56.79140903+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T19:37:56.798693744+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"etkbn0o2","id":"1(@)"} +{"time":"2025-09-16T19:37:57.258519078+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"etkbn0o2","id":"1(@)"} +{"time":"2025-09-16T19:39:20.040420906+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T19:39:20.040502677+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T19:39:20.040630162+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T19:39:20.040524956+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T19:39:20.041905463+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3884253-3884884-4178420690/socket","Net":"unix"}} +{"time":"2025-09-16T19:39:21.075505167+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T19:39:21.075562608+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T19:39:21.075598649+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_193756-etkbn0o2/logs/debug-internal.log b/wandb/run-20250916_193756-etkbn0o2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..13663455b4074c528574974bb6bab4e890ac76aa --- /dev/null +++ b/wandb/run-20250916_193756-etkbn0o2/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T19:37:56.798970023+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T19:37:56.821633166+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T19:37:57.25840013+08:00","level":"INFO","msg":"stream: created new stream","id":"etkbn0o2"} +{"time":"2025-09-16T19:37:57.258504602+08:00","level":"INFO","msg":"stream: started","id":"etkbn0o2"} +{"time":"2025-09-16T19:37:57.258535951+08:00","level":"INFO","msg":"writer: started","stream_id":"etkbn0o2"} +{"time":"2025-09-16T19:37:57.258551618+08:00","level":"INFO","msg":"sender: started","stream_id":"etkbn0o2"} +{"time":"2025-09-16T19:37:57.258580383+08:00","level":"INFO","msg":"handler: started","stream_id":"etkbn0o2"} +{"time":"2025-09-16T19:39:20.04052383+08:00","level":"INFO","msg":"stream: closing","id":"etkbn0o2"} +{"time":"2025-09-16T19:39:20.793397488+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T19:39:21.073900443+08:00","level":"INFO","msg":"handler: closed","stream_id":"etkbn0o2"} +{"time":"2025-09-16T19:39:21.074435997+08:00","level":"INFO","msg":"sender: closed","stream_id":"etkbn0o2"} +{"time":"2025-09-16T19:39:21.074481351+08:00","level":"INFO","msg":"stream: closed","id":"etkbn0o2"} diff --git a/wandb/run-20250916_193756-etkbn0o2/logs/debug.log b/wandb/run-20250916_193756-etkbn0o2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..dac09a4e99aa426b9826e1923661cab13cc00bf5 --- /dev/null +++ b/wandb/run-20250916_193756-etkbn0o2/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 19:37:56,571 INFO MainThread:3884253 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 19:37:56,571 INFO MainThread:3884253 [wandb_setup.py:_flush():81] Configure stats pid to 3884253 +2025-09-16 19:37:56,571 INFO MainThread:3884253 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_193756-etkbn0o2/logs/debug.log +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_193756-etkbn0o2/logs/debug-internal.log +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_init.py:init():813] calling init triggers +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 19:37:56,572 INFO MainThread:3884253 [wandb_init.py:init():854] starting backend +2025-09-16 19:37:56,791 INFO MainThread:3884253 [wandb_init.py:init():857] sending inform_init request +2025-09-16 19:37:56,793 INFO MainThread:3884253 [wandb_init.py:init():865] backend started and connected +2025-09-16 19:37:56,795 INFO MainThread:3884253 [wandb_init.py:init():936] updated telemetry +2025-09-16 19:37:56,802 INFO MainThread:3884253 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 19:37:57,564 INFO MainThread:3884253 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 19:37:57,740 INFO MainThread:3884253 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 19:37:57,741 INFO MainThread:3884253 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 19:37:57,742 INFO MainThread:3884253 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 19:37:57,742 INFO MainThread:3884253 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 19:37:57,745 INFO MainThread:3884253 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 19:39:05,662 INFO MainThread:3884253 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 19:39:20,040 INFO wandb-AsyncioManager-main:3884253 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 19:39:20,040 INFO wandb-AsyncioManager-main:3884253 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_193756-etkbn0o2/run-etkbn0o2.wandb b/wandb/run-20250916_193756-etkbn0o2/run-etkbn0o2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e4a8652a5677ca869a847fee98ee59c17de53854 Binary files /dev/null and b/wandb/run-20250916_193756-etkbn0o2/run-etkbn0o2.wandb differ diff --git a/wandb/run-20250916_194050-886jzssh/files/output.log b/wandb/run-20250916_194050-886jzssh/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8cad20329f3801e96986170ba9fc084ae1778c92 --- /dev/null +++ b/wandb/run-20250916_194050-886jzssh/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:52<11:55:41, 0.04it/s, v_num=zssh] diff --git a/wandb/run-20250916_194050-886jzssh/files/requirements.txt b/wandb/run-20250916_194050-886jzssh/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_194050-886jzssh/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_194050-886jzssh/files/wandb-metadata.json b/wandb/run-20250916_194050-886jzssh/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bb67e7c6239fe1553373495bc86b834fb13249d2 --- /dev/null +++ b/wandb/run-20250916_194050-886jzssh/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T11:40:50.664220Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576975765504" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "uut0zjus7bmehz5ie6y3kx12dvxny1yl" +} \ No newline at end of file diff --git a/wandb/run-20250916_194050-886jzssh/logs/debug-core.log b/wandb/run-20250916_194050-886jzssh/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..12518840898450a986cd78f4fc91acc985a047d4 --- /dev/null +++ b/wandb/run-20250916_194050-886jzssh/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T19:40:50.700304895+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp5md39evi/port-3892570.txt","pid":3892570,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T19:40:50.700561987+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T19:40:50.701222883+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3892570} +{"time":"2025-09-16T19:40:50.701183903+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3892570-3893247-3854762527/socket","Net":"unix"}} +{"time":"2025-09-16T19:40:50.889342428+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T19:40:50.897218086+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"886jzssh","id":"1(@)"} +{"time":"2025-09-16T19:40:51.34536086+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"886jzssh","id":"1(@)"} +{"time":"2025-09-16T19:44:58.065784158+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_194050-886jzssh/logs/debug-internal.log b/wandb/run-20250916_194050-886jzssh/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..fb6ba58fb2a59f59d08aabcd1c941c12a53d9654 --- /dev/null +++ b/wandb/run-20250916_194050-886jzssh/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T19:40:50.897348817+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T19:40:50.91208468+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T19:40:51.345292727+08:00","level":"INFO","msg":"stream: created new stream","id":"886jzssh"} +{"time":"2025-09-16T19:40:51.345352485+08:00","level":"INFO","msg":"stream: started","id":"886jzssh"} +{"time":"2025-09-16T19:40:51.345391288+08:00","level":"INFO","msg":"handler: started","stream_id":"886jzssh"} +{"time":"2025-09-16T19:40:51.345393608+08:00","level":"INFO","msg":"writer: started","stream_id":"886jzssh"} +{"time":"2025-09-16T19:40:51.34546845+08:00","level":"INFO","msg":"sender: started","stream_id":"886jzssh"} diff --git a/wandb/run-20250916_194050-886jzssh/logs/debug.log b/wandb/run-20250916_194050-886jzssh/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..faf9b28a6651a6c85b40ba562273125250778f44 --- /dev/null +++ b/wandb/run-20250916_194050-886jzssh/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 19:40:50,668 INFO MainThread:3892570 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 19:40:50,668 INFO MainThread:3892570 [wandb_setup.py:_flush():81] Configure stats pid to 3892570 +2025-09-16 19:40:50,668 INFO MainThread:3892570 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 19:40:50,668 INFO MainThread:3892570 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 19:40:50,668 INFO MainThread:3892570 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 19:40:50,668 INFO MainThread:3892570 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_194050-886jzssh/logs/debug.log +2025-09-16 19:40:50,669 INFO MainThread:3892570 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_194050-886jzssh/logs/debug-internal.log +2025-09-16 19:40:50,669 INFO MainThread:3892570 [wandb_init.py:init():813] calling init triggers +2025-09-16 19:40:50,669 INFO MainThread:3892570 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 19:40:50,669 INFO MainThread:3892570 [wandb_init.py:init():854] starting backend +2025-09-16 19:40:50,889 INFO MainThread:3892570 [wandb_init.py:init():857] sending inform_init request +2025-09-16 19:40:50,894 INFO MainThread:3892570 [wandb_init.py:init():865] backend started and connected +2025-09-16 19:40:50,897 INFO MainThread:3892570 [wandb_init.py:init():936] updated telemetry +2025-09-16 19:40:50,908 INFO MainThread:3892570 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 19:40:51,891 INFO MainThread:3892570 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 19:40:52,075 INFO MainThread:3892570 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 19:40:52,075 INFO MainThread:3892570 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 19:40:52,075 INFO MainThread:3892570 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 19:40:52,075 INFO MainThread:3892570 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 19:40:52,078 INFO MainThread:3892570 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 19:42:00,719 INFO MainThread:3892570 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_194050-886jzssh/run-886jzssh.wandb b/wandb/run-20250916_194050-886jzssh/run-886jzssh.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_194407-0cobmvm9/files/config.yaml b/wandb/run-20250916_194407-0cobmvm9/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..33b415b4196926942475d98dd0d2f894a1890cb7 --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + qkd7kcyarvosrwikgd5j2jsyz0toj4g2: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3576976424960" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T11:44:07.701330Z" + writerId: qkd7kcyarvosrwikgd5j2jsyz0toj4g2 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_194407-0cobmvm9/files/output.log b/wandb/run-20250916_194407-0cobmvm9/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_194407-0cobmvm9/files/requirements.txt b/wandb/run-20250916_194407-0cobmvm9/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_194407-0cobmvm9/files/wandb-metadata.json b/wandb/run-20250916_194407-0cobmvm9/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..50b631f93a497e55b7943dd1d15df78f062414ae --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T11:44:07.701330Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576976424960" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "qkd7kcyarvosrwikgd5j2jsyz0toj4g2" +} \ No newline at end of file diff --git a/wandb/run-20250916_194407-0cobmvm9/files/wandb-summary.json b/wandb/run-20250916_194407-0cobmvm9/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..99bd83bf5edb6669d823b5cece522f21f090dd47 --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":32,"_wandb":{"runtime":32}} \ No newline at end of file diff --git a/wandb/run-20250916_194407-0cobmvm9/logs/debug-core.log b/wandb/run-20250916_194407-0cobmvm9/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..a2d0c3e1c01bc45ec2639464f64853a9fb15173e --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-16T19:44:07.735608677+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmptx6x5srm/port-3901374.txt","pid":3901374,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T19:44:07.735790512+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T19:44:07.736312665+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3901374} +{"time":"2025-09-16T19:44:07.736307486+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3901374-3902094-1849290847/socket","Net":"unix"}} +{"time":"2025-09-16T19:44:07.925378521+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T19:44:07.932937018+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"0cobmvm9","id":"1(@)"} +{"time":"2025-09-16T19:44:08.394965675+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"0cobmvm9","id":"1(@)"} +{"time":"2025-09-16T19:44:41.504227446+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T19:44:41.504346117+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T19:44:41.504371223+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T19:44:41.504422978+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T19:44:41.504683484+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3901374-3902094-1849290847/socket","Net":"unix"}} +{"time":"2025-09-16T19:44:42.141301264+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_194407-0cobmvm9/logs/debug-internal.log b/wandb/run-20250916_194407-0cobmvm9/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..31dda00c7ced7b641ed59d953acb2629794c1bc3 --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T19:44:07.933083398+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T19:44:07.957830246+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T19:44:08.394878943+08:00","level":"INFO","msg":"stream: created new stream","id":"0cobmvm9"} +{"time":"2025-09-16T19:44:08.394955453+08:00","level":"INFO","msg":"stream: started","id":"0cobmvm9"} +{"time":"2025-09-16T19:44:08.394980771+08:00","level":"INFO","msg":"sender: started","stream_id":"0cobmvm9"} +{"time":"2025-09-16T19:44:08.394985999+08:00","level":"INFO","msg":"writer: started","stream_id":"0cobmvm9"} +{"time":"2025-09-16T19:44:08.39502389+08:00","level":"INFO","msg":"handler: started","stream_id":"0cobmvm9"} +{"time":"2025-09-16T19:44:41.504350363+08:00","level":"INFO","msg":"stream: closing","id":"0cobmvm9"} diff --git a/wandb/run-20250916_194407-0cobmvm9/logs/debug.log b/wandb/run-20250916_194407-0cobmvm9/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7d310e97845f5778ca1f45df8138c49bcc6827b8 --- /dev/null +++ b/wandb/run-20250916_194407-0cobmvm9/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 19:44:07,705 INFO MainThread:3901374 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 19:44:07,705 INFO MainThread:3901374 [wandb_setup.py:_flush():81] Configure stats pid to 3901374 +2025-09-16 19:44:07,705 INFO MainThread:3901374 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 19:44:07,705 INFO MainThread:3901374 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 19:44:07,705 INFO MainThread:3901374 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 19:44:07,705 INFO MainThread:3901374 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_194407-0cobmvm9/logs/debug.log +2025-09-16 19:44:07,706 INFO MainThread:3901374 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_194407-0cobmvm9/logs/debug-internal.log +2025-09-16 19:44:07,706 INFO MainThread:3901374 [wandb_init.py:init():813] calling init triggers +2025-09-16 19:44:07,706 INFO MainThread:3901374 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 19:44:07,706 INFO MainThread:3901374 [wandb_init.py:init():854] starting backend +2025-09-16 19:44:07,925 INFO MainThread:3901374 [wandb_init.py:init():857] sending inform_init request +2025-09-16 19:44:07,930 INFO MainThread:3901374 [wandb_init.py:init():865] backend started and connected +2025-09-16 19:44:07,933 INFO MainThread:3901374 [wandb_init.py:init():936] updated telemetry +2025-09-16 19:44:07,943 INFO MainThread:3901374 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 19:44:08,700 INFO MainThread:3901374 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 19:44:08,868 INFO MainThread:3901374 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 19:44:08,868 INFO MainThread:3901374 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 19:44:08,868 INFO MainThread:3901374 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 19:44:08,868 INFO MainThread:3901374 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 19:44:08,871 INFO MainThread:3901374 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 19:44:41,504 INFO wandb-AsyncioManager-main:3901374 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 19:44:41,504 INFO wandb-AsyncioManager-main:3901374 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_194407-0cobmvm9/run-0cobmvm9.wandb b/wandb/run-20250916_194407-0cobmvm9/run-0cobmvm9.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_194536-y22mzabg/files/output.log b/wandb/run-20250916_194536-y22mzabg/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..686ce80871163f08f1ede6659cd352404397ae94 --- /dev/null +++ b/wandb/run-20250916_194536-y22mzabg/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:51<11:54:55, 0.04it/s, v_num=zabg] diff --git a/wandb/run-20250916_194536-y22mzabg/files/requirements.txt b/wandb/run-20250916_194536-y22mzabg/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_194536-y22mzabg/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_194536-y22mzabg/files/wandb-metadata.json b/wandb/run-20250916_194536-y22mzabg/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8ef410e591d4d994a1b09d9e6160643d83fa5877 --- /dev/null +++ b/wandb/run-20250916_194536-y22mzabg/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T11:45:36.845751Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576976605184" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "67i2jgqxc1wc3gqrjnyco9cba01cnipo" +} \ No newline at end of file diff --git a/wandb/run-20250916_194536-y22mzabg/logs/debug-core.log b/wandb/run-20250916_194536-y22mzabg/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..9392dda3601c41acba0b520458a73a43b8070cd7 --- /dev/null +++ b/wandb/run-20250916_194536-y22mzabg/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T19:45:36.880978247+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp4afdwbew/port-3905526.txt","pid":3905526,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T19:45:36.881975354+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T19:45:36.882662618+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3905526} +{"time":"2025-09-16T19:45:36.882640627+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3905526-3906849-4047425087/socket","Net":"unix"}} +{"time":"2025-09-16T19:45:37.070360214+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T19:45:37.07937215+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"y22mzabg","id":"1(@)"} +{"time":"2025-09-16T19:45:37.567238464+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"y22mzabg","id":"1(@)"} +{"time":"2025-09-16T19:48:21.932161094+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_194536-y22mzabg/logs/debug-internal.log b/wandb/run-20250916_194536-y22mzabg/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d7f1b20eb7ded083fe231a55b26145040bd1cd6b --- /dev/null +++ b/wandb/run-20250916_194536-y22mzabg/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T19:45:37.079525881+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T19:45:37.09885148+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T19:45:37.567163125+08:00","level":"INFO","msg":"stream: created new stream","id":"y22mzabg"} +{"time":"2025-09-16T19:45:37.567232693+08:00","level":"INFO","msg":"stream: started","id":"y22mzabg"} +{"time":"2025-09-16T19:45:37.567270309+08:00","level":"INFO","msg":"writer: started","stream_id":"y22mzabg"} +{"time":"2025-09-16T19:45:37.567285659+08:00","level":"INFO","msg":"handler: started","stream_id":"y22mzabg"} +{"time":"2025-09-16T19:45:37.567318399+08:00","level":"INFO","msg":"sender: started","stream_id":"y22mzabg"} diff --git a/wandb/run-20250916_194536-y22mzabg/logs/debug.log b/wandb/run-20250916_194536-y22mzabg/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..0a215f831a3de4f6e71dca95509dbc67f8c17ab6 --- /dev/null +++ b/wandb/run-20250916_194536-y22mzabg/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 19:45:36,849 INFO MainThread:3905526 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 19:45:36,849 INFO MainThread:3905526 [wandb_setup.py:_flush():81] Configure stats pid to 3905526 +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_194536-y22mzabg/logs/debug.log +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_194536-y22mzabg/logs/debug-internal.log +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_init.py:init():813] calling init triggers +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 19:45:36,850 INFO MainThread:3905526 [wandb_init.py:init():854] starting backend +2025-09-16 19:45:37,070 INFO MainThread:3905526 [wandb_init.py:init():857] sending inform_init request +2025-09-16 19:45:37,074 INFO MainThread:3905526 [wandb_init.py:init():865] backend started and connected +2025-09-16 19:45:37,075 INFO MainThread:3905526 [wandb_init.py:init():936] updated telemetry +2025-09-16 19:45:37,084 INFO MainThread:3905526 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 19:45:37,851 INFO MainThread:3905526 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 19:45:37,965 INFO MainThread:3905526 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 19:45:37,965 INFO MainThread:3905526 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 19:45:37,966 INFO MainThread:3905526 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 19:45:37,966 INFO MainThread:3905526 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 19:45:37,967 INFO MainThread:3905526 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 19:46:46,828 INFO MainThread:3905526 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_194536-y22mzabg/run-y22mzabg.wandb b/wandb/run-20250916_194536-y22mzabg/run-y22mzabg.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_195304-sb30r8u1/files/config.yaml b/wandb/run-20250916_195304-sb30r8u1/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1914cee02e2a5c7124deb47491764584b3b66036 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + xwscwet5r6niggl4h3d8lhhnwlg25msn: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3576977317888" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T11:53:04.358783Z" + writerId: xwscwet5r6niggl4h3d8lhhnwlg25msn + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_195304-sb30r8u1/files/output.log b/wandb/run-20250916_195304-sb30r8u1/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..cea5fd0d19d5e960248a5f194f6732d73e3c02f7 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/files/output.log @@ -0,0 +1,185 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 570, in training_step + actor_loss, actor_log = self.actor_loss(batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 650, in + self.actor_loss = lambda batch: self.awr_loss(**batch) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 654, in awr_loss + log_prob = self.actor.get_logsum_prob(observation, action) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 136, in get_logsum_prob + if not self.model.gradient_checkpointing: + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 860, in __getattr__ + return getattr(self.base_model, name) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 386, in __getattr__ + return getattr(self.model, name) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ + raise AttributeError( +AttributeError: 'Qwen3ForCausalLM' object has no attribute 'gradient_checkpointing' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 856, in __getattr__ +[rank0]: return super().__getattr__(name) # defer to nn.Module's logic +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'PeftModelForCausalLM' object has no attribute 'gradient_checkpointing' + +[rank0]: During handling of the above exception, another exception occurred: + +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 382, in __getattr__ +[rank0]: return super().__getattr__(name) # defer to nn.Module's logic +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'LoraModel' object has no attribute 'gradient_checkpointing' + +[rank0]: During handling of the above exception, another exception occurred: + +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 570, in training_step +[rank0]: actor_loss, actor_log = self.actor_loss(batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 650, in +[rank0]: self.actor_loss = lambda batch: self.awr_loss(**batch) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 654, in awr_loss +[rank0]: log_prob = self.actor.get_logsum_prob(observation, action) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 136, in get_logsum_prob +[rank0]: if not self.model.gradient_checkpointing: +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/peft_model.py", line 860, in __getattr__ +[rank0]: return getattr(self.base_model, name) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/peft/tuners/lora/model.py", line 386, in __getattr__ +[rank0]: return getattr(self.model, name) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'Qwen3ForCausalLM' object has no attribute 'gradient_checkpointing' diff --git a/wandb/run-20250916_195304-sb30r8u1/files/requirements.txt b/wandb/run-20250916_195304-sb30r8u1/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_195304-sb30r8u1/files/wandb-metadata.json b/wandb/run-20250916_195304-sb30r8u1/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..931d629198759a58686827da6644b1227d1ebe42 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T11:53:04.358783Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576977317888" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "xwscwet5r6niggl4h3d8lhhnwlg25msn" +} \ No newline at end of file diff --git a/wandb/run-20250916_195304-sb30r8u1/files/wandb-summary.json b/wandb/run-20250916_195304-sb30r8u1/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..6f3f1a3ec425c7e952261589df6ff9be78972f24 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":72},"_runtime":72} \ No newline at end of file diff --git a/wandb/run-20250916_195304-sb30r8u1/logs/debug-core.log b/wandb/run-20250916_195304-sb30r8u1/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..a06137a147935a11cb2f63b4513a65c524b40f4e --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T19:53:04.393574137+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp9u_mr0he/port-3923856.txt","pid":3923856,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T19:53:04.393863741+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T19:53:04.394397785+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3923856} +{"time":"2025-09-16T19:53:04.39437041+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3923856-3924479-1367472272/socket","Net":"unix"}} +{"time":"2025-09-16T19:53:04.582693455+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T19:53:04.593342921+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"sb30r8u1","id":"1(@)"} +{"time":"2025-09-16T19:53:05.048202889+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"sb30r8u1","id":"1(@)"} +{"time":"2025-09-16T19:54:17.788720884+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T19:54:17.788796723+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T19:54:17.788866124+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T19:54:17.788812379+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T19:54:17.789078379+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-3923856-3924479-1367472272/socket","Net":"unix"}} +{"time":"2025-09-16T19:54:19.401615491+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T19:54:19.401672255+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T19:54:19.401708854+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_195304-sb30r8u1/logs/debug-internal.log b/wandb/run-20250916_195304-sb30r8u1/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..ff48f54ffeae8a733b84f1c3881c505ddb05b346 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T19:53:04.593528042+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T19:53:04.613138591+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T19:53:05.048093397+08:00","level":"INFO","msg":"stream: created new stream","id":"sb30r8u1"} +{"time":"2025-09-16T19:53:05.048169586+08:00","level":"INFO","msg":"stream: started","id":"sb30r8u1"} +{"time":"2025-09-16T19:53:05.048194232+08:00","level":"INFO","msg":"sender: started","stream_id":"sb30r8u1"} +{"time":"2025-09-16T19:53:05.048181306+08:00","level":"INFO","msg":"handler: started","stream_id":"sb30r8u1"} +{"time":"2025-09-16T19:53:05.048205812+08:00","level":"INFO","msg":"writer: started","stream_id":"sb30r8u1"} +{"time":"2025-09-16T19:54:17.788783414+08:00","level":"INFO","msg":"stream: closing","id":"sb30r8u1"} +{"time":"2025-09-16T19:54:19.102401701+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T19:54:19.400541019+08:00","level":"INFO","msg":"handler: closed","stream_id":"sb30r8u1"} +{"time":"2025-09-16T19:54:19.401058763+08:00","level":"INFO","msg":"sender: closed","stream_id":"sb30r8u1"} +{"time":"2025-09-16T19:54:19.401106827+08:00","level":"INFO","msg":"stream: closed","id":"sb30r8u1"} diff --git a/wandb/run-20250916_195304-sb30r8u1/logs/debug.log b/wandb/run-20250916_195304-sb30r8u1/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a327fa518eeb8cb3af0c89a77e33bcbec05b1124 --- /dev/null +++ b/wandb/run-20250916_195304-sb30r8u1/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 19:53:04,362 INFO MainThread:3923856 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 19:53:04,362 INFO MainThread:3923856 [wandb_setup.py:_flush():81] Configure stats pid to 3923856 +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_195304-sb30r8u1/logs/debug.log +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_195304-sb30r8u1/logs/debug-internal.log +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_init.py:init():813] calling init triggers +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 19:53:04,363 INFO MainThread:3923856 [wandb_init.py:init():854] starting backend +2025-09-16 19:53:04,582 INFO MainThread:3923856 [wandb_init.py:init():857] sending inform_init request +2025-09-16 19:53:04,587 INFO MainThread:3923856 [wandb_init.py:init():865] backend started and connected +2025-09-16 19:53:04,589 INFO MainThread:3923856 [wandb_init.py:init():936] updated telemetry +2025-09-16 19:53:04,600 INFO MainThread:3923856 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 19:53:05,357 INFO MainThread:3923856 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 19:53:05,523 INFO MainThread:3923856 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 19:53:05,523 INFO MainThread:3923856 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 19:53:05,523 INFO MainThread:3923856 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 19:53:05,523 INFO MainThread:3923856 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 19:53:05,526 INFO MainThread:3923856 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 19:54:13,223 INFO MainThread:3923856 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 19:54:17,788 INFO wandb-AsyncioManager-main:3923856 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 19:54:17,789 INFO wandb-AsyncioManager-main:3923856 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_195304-sb30r8u1/run-sb30r8u1.wandb b/wandb/run-20250916_195304-sb30r8u1/run-sb30r8u1.wandb new file mode 100644 index 0000000000000000000000000000000000000000..473fa6bea517ad1b68ad679a3a47f82652ff47ba Binary files /dev/null and b/wandb/run-20250916_195304-sb30r8u1/run-sb30r8u1.wandb differ diff --git a/wandb/run-20250916_195748-gs054p58/files/output.log b/wandb/run-20250916_195748-gs054p58/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4f21820051f3b6c74ec66ead7f9325bda2b0da8b --- /dev/null +++ b/wandb/run-20250916_195748-gs054p58/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:52<11:55:19, 0.04it/s, v_num=4p58] diff --git a/wandb/run-20250916_195748-gs054p58/files/requirements.txt b/wandb/run-20250916_195748-gs054p58/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_195748-gs054p58/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_195748-gs054p58/files/wandb-metadata.json b/wandb/run-20250916_195748-gs054p58/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..d18a1dc0cd5561a6d8168f82a7c85d79dec98162 --- /dev/null +++ b/wandb/run-20250916_195748-gs054p58/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T11:57:48.185148Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576977723392" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "2zegl63jbwhphyg1hyyqe4q99a87fvvj" +} \ No newline at end of file diff --git a/wandb/run-20250916_195748-gs054p58/logs/debug-core.log b/wandb/run-20250916_195748-gs054p58/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..0abd8ccca592c790acd5fd34b3ee5581a7741ab3 --- /dev/null +++ b/wandb/run-20250916_195748-gs054p58/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T19:57:48.220116406+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpv8z9k651/port-3935424.txt","pid":3935424,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T19:57:48.220409402+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T19:57:48.220964343+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3935424} +{"time":"2025-09-16T19:57:48.220974678+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3935424-3936159-3864044177/socket","Net":"unix"}} +{"time":"2025-09-16T19:57:48.409531981+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T19:57:48.419719044+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"gs054p58","id":"1(@)"} +{"time":"2025-09-16T19:57:48.859004983+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"gs054p58","id":"1(@)"} +{"time":"2025-09-16T20:00:33.182862741+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_195748-gs054p58/logs/debug-internal.log b/wandb/run-20250916_195748-gs054p58/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..580177f4dedb869f82febacc0b6e26b85c0979e9 --- /dev/null +++ b/wandb/run-20250916_195748-gs054p58/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T19:57:48.41987754+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T19:57:48.435892875+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T19:57:48.85892663+08:00","level":"INFO","msg":"stream: created new stream","id":"gs054p58"} +{"time":"2025-09-16T19:57:48.858995845+08:00","level":"INFO","msg":"stream: started","id":"gs054p58"} +{"time":"2025-09-16T19:57:48.859033483+08:00","level":"INFO","msg":"sender: started","stream_id":"gs054p58"} +{"time":"2025-09-16T19:57:48.859081435+08:00","level":"INFO","msg":"writer: started","stream_id":"gs054p58"} +{"time":"2025-09-16T19:57:48.859021695+08:00","level":"INFO","msg":"handler: started","stream_id":"gs054p58"} diff --git a/wandb/run-20250916_195748-gs054p58/logs/debug.log b/wandb/run-20250916_195748-gs054p58/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..87c66dd5cdcd8ee7161c93f4bb91632fe40316d1 --- /dev/null +++ b/wandb/run-20250916_195748-gs054p58/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 19:57:48,188 INFO MainThread:3935424 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_setup.py:_flush():81] Configure stats pid to 3935424 +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_195748-gs054p58/logs/debug.log +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_195748-gs054p58/logs/debug-internal.log +2025-09-16 19:57:48,189 INFO MainThread:3935424 [wandb_init.py:init():813] calling init triggers +2025-09-16 19:57:48,190 INFO MainThread:3935424 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 19:57:48,190 INFO MainThread:3935424 [wandb_init.py:init():854] starting backend +2025-09-16 19:57:48,409 INFO MainThread:3935424 [wandb_init.py:init():857] sending inform_init request +2025-09-16 19:57:48,414 INFO MainThread:3935424 [wandb_init.py:init():865] backend started and connected +2025-09-16 19:57:48,416 INFO MainThread:3935424 [wandb_init.py:init():936] updated telemetry +2025-09-16 19:57:48,428 INFO MainThread:3935424 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 19:57:49,194 INFO MainThread:3935424 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 19:57:49,357 INFO MainThread:3935424 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 19:57:49,357 INFO MainThread:3935424 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 19:57:49,358 INFO MainThread:3935424 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 19:57:49,358 INFO MainThread:3935424 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 19:57:49,361 INFO MainThread:3935424 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 19:58:58,057 INFO MainThread:3935424 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_195748-gs054p58/run-gs054p58.wandb b/wandb/run-20250916_195748-gs054p58/run-gs054p58.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_200311-4usj50tm/files/output.log b/wandb/run-20250916_200311-4usj50tm/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..eff0178a55eb24ddaccda50635697133a720e268 --- /dev/null +++ b/wandb/run-20250916_200311-4usj50tm/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 3310600 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:52<12:06:27, 0.04it/s, v_num=50tm] diff --git a/wandb/run-20250916_200311-4usj50tm/files/requirements.txt b/wandb/run-20250916_200311-4usj50tm/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_200311-4usj50tm/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_200311-4usj50tm/files/wandb-metadata.json b/wandb/run-20250916_200311-4usj50tm/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..a7336230f226e79cef416307fbd72afbeb995719 --- /dev/null +++ b/wandb/run-20250916_200311-4usj50tm/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T12:03:11.314830Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576953368576" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "a43fn2q00u3qrpyiolxrgkm7igzcrx2e" +} \ No newline at end of file diff --git a/wandb/run-20250916_200311-4usj50tm/logs/debug-core.log b/wandb/run-20250916_200311-4usj50tm/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..3c9a36a5dca5e910caf245daafb3ee6db53b3fff --- /dev/null +++ b/wandb/run-20250916_200311-4usj50tm/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T20:03:11.37672236+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpwhyxmvms/port-3947961.txt","pid":3947961,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T20:03:11.376903441+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T20:03:11.37747278+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3947961} +{"time":"2025-09-16T20:03:11.377481463+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3947961-3949369-4286855109/socket","Net":"unix"}} +{"time":"2025-09-16T20:03:11.529821179+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T20:03:11.539027229+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"4usj50tm","id":"1(@)"} +{"time":"2025-09-16T20:03:12.037677153+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"4usj50tm","id":"1(@)"} +{"time":"2025-09-16T20:06:42.959465806+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_200311-4usj50tm/logs/debug-internal.log b/wandb/run-20250916_200311-4usj50tm/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..06d814f52bc1b69d37cf8da16796fe09c462fa50 --- /dev/null +++ b/wandb/run-20250916_200311-4usj50tm/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T20:03:11.539149622+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T20:03:11.565635989+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T20:03:12.037564718+08:00","level":"INFO","msg":"stream: created new stream","id":"4usj50tm"} +{"time":"2025-09-16T20:03:12.03766216+08:00","level":"INFO","msg":"stream: started","id":"4usj50tm"} +{"time":"2025-09-16T20:03:12.037709443+08:00","level":"INFO","msg":"writer: started","stream_id":"4usj50tm"} +{"time":"2025-09-16T20:03:12.037746423+08:00","level":"INFO","msg":"sender: started","stream_id":"4usj50tm"} +{"time":"2025-09-16T20:03:12.03771463+08:00","level":"INFO","msg":"handler: started","stream_id":"4usj50tm"} diff --git a/wandb/run-20250916_200311-4usj50tm/logs/debug.log b/wandb/run-20250916_200311-4usj50tm/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..008d3ec95212c3c1c7953a1629d1d01d808d4ec3 --- /dev/null +++ b/wandb/run-20250916_200311-4usj50tm/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_setup.py:_flush():81] Configure stats pid to 3947961 +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_200311-4usj50tm/logs/debug.log +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_200311-4usj50tm/logs/debug-internal.log +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_init.py:init():813] calling init triggers +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 20:03:11,318 INFO MainThread:3947961 [wandb_init.py:init():854] starting backend +2025-09-16 20:03:11,530 INFO MainThread:3947961 [wandb_init.py:init():857] sending inform_init request +2025-09-16 20:03:11,534 INFO MainThread:3947961 [wandb_init.py:init():865] backend started and connected +2025-09-16 20:03:11,537 INFO MainThread:3947961 [wandb_init.py:init():936] updated telemetry +2025-09-16 20:03:11,550 INFO MainThread:3947961 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 20:03:12,479 INFO MainThread:3947961 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 20:03:12,646 INFO MainThread:3947961 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 20:03:12,646 INFO MainThread:3947961 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 20:03:12,647 INFO MainThread:3947961 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 20:03:12,647 INFO MainThread:3947961 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 20:03:12,651 INFO MainThread:3947961 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 20:04:57,460 INFO MainThread:3947961 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_200311-4usj50tm/run-4usj50tm.wandb b/wandb/run-20250916_200311-4usj50tm/run-4usj50tm.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_200723-j3kg70g4/files/output.log b/wandb/run-20250916_200723-j3kg70g4/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..09c929d61a78fd024860386ffd0e6e9c2344cecf --- /dev/null +++ b/wandb/run-20250916_200723-j3kg70g4/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 3310600 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:51<11:50:29, 0.04it/s, v_num=70g4] diff --git a/wandb/run-20250916_200723-j3kg70g4/files/requirements.txt b/wandb/run-20250916_200723-j3kg70g4/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_200723-j3kg70g4/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_200723-j3kg70g4/files/wandb-metadata.json b/wandb/run-20250916_200723-j3kg70g4/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..05ae2f61df8da8dd811ee8c1e966a1ed5c0981c0 --- /dev/null +++ b/wandb/run-20250916_200723-j3kg70g4/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T12:07:23.099727Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576953810944" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "adjs8b21dmfah9b6lujqn92yklczqsld" +} \ No newline at end of file diff --git a/wandb/run-20250916_200723-j3kg70g4/logs/debug-core.log b/wandb/run-20250916_200723-j3kg70g4/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..93e0ac86d1430b2edb3f5e74090452dc4b50869b --- /dev/null +++ b/wandb/run-20250916_200723-j3kg70g4/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T20:07:23.125891687+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpn3ilui1i/port-3959402.txt","pid":3959402,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T20:07:23.126097595+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T20:07:23.126889073+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3959402} +{"time":"2025-09-16T20:07:23.126892141+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3959402-3960088-2906741653/socket","Net":"unix"}} +{"time":"2025-09-16T20:07:23.314728456+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T20:07:23.32239328+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"j3kg70g4","id":"1(@)"} +{"time":"2025-09-16T20:07:23.77706323+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"j3kg70g4","id":"1(@)"} +{"time":"2025-09-16T20:10:20.125183551+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_200723-j3kg70g4/logs/debug-internal.log b/wandb/run-20250916_200723-j3kg70g4/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a0e40040bb49824f38b0af445eb860749ff71c6c --- /dev/null +++ b/wandb/run-20250916_200723-j3kg70g4/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T20:07:23.322535148+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T20:07:23.338282387+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T20:07:23.776858952+08:00","level":"INFO","msg":"stream: created new stream","id":"j3kg70g4"} +{"time":"2025-09-16T20:07:23.777046241+08:00","level":"INFO","msg":"stream: started","id":"j3kg70g4"} +{"time":"2025-09-16T20:07:23.777051765+08:00","level":"INFO","msg":"writer: started","stream_id":"j3kg70g4"} +{"time":"2025-09-16T20:07:23.777078937+08:00","level":"INFO","msg":"handler: started","stream_id":"j3kg70g4"} +{"time":"2025-09-16T20:07:23.77711639+08:00","level":"INFO","msg":"sender: started","stream_id":"j3kg70g4"} diff --git a/wandb/run-20250916_200723-j3kg70g4/logs/debug.log b/wandb/run-20250916_200723-j3kg70g4/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..ae0bd112e5de6c5349c59e0684d9fcda72e6a74a --- /dev/null +++ b/wandb/run-20250916_200723-j3kg70g4/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_setup.py:_flush():81] Configure stats pid to 3959402 +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_200723-j3kg70g4/logs/debug.log +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_200723-j3kg70g4/logs/debug-internal.log +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_init.py:init():813] calling init triggers +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 20:07:23,102 INFO MainThread:3959402 [wandb_init.py:init():854] starting backend +2025-09-16 20:07:23,314 INFO MainThread:3959402 [wandb_init.py:init():857] sending inform_init request +2025-09-16 20:07:23,317 INFO MainThread:3959402 [wandb_init.py:init():865] backend started and connected +2025-09-16 20:07:23,318 INFO MainThread:3959402 [wandb_init.py:init():936] updated telemetry +2025-09-16 20:07:23,325 INFO MainThread:3959402 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 20:07:24,428 INFO MainThread:3959402 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 20:07:24,611 INFO MainThread:3959402 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 20:07:24,612 INFO MainThread:3959402 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 20:07:24,612 INFO MainThread:3959402 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 20:07:24,612 INFO MainThread:3959402 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 20:07:24,614 INFO MainThread:3959402 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 20:08:45,004 INFO MainThread:3959402 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_200723-j3kg70g4/run-j3kg70g4.wandb b/wandb/run-20250916_200723-j3kg70g4/run-j3kg70g4.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_201356-7v3q4wn2/files/output.log b/wandb/run-20250916_201356-7v3q4wn2/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..49e0e818fab3b2b973fe7f623b5466ec9f6399a6 --- /dev/null +++ b/wandb/run-20250916_201356-7v3q4wn2/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 3310600 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:52<12:04:27, 0.04it/s, v_num=4wn2] diff --git a/wandb/run-20250916_201356-7v3q4wn2/files/requirements.txt b/wandb/run-20250916_201356-7v3q4wn2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_201356-7v3q4wn2/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_201356-7v3q4wn2/files/wandb-metadata.json b/wandb/run-20250916_201356-7v3q4wn2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..9dc2c8a2a1610b928879ad55eebbcff340940c81 --- /dev/null +++ b/wandb/run-20250916_201356-7v3q4wn2/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T12:13:56.420459Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576954499072" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "sz9qzyfm2irogf5wi85oqwvcjdrxw7rg" +} \ No newline at end of file diff --git a/wandb/run-20250916_201356-7v3q4wn2/logs/debug-core.log b/wandb/run-20250916_201356-7v3q4wn2/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..5ca7c25c6cd747a9b53808cc69df7669d9696429 --- /dev/null +++ b/wandb/run-20250916_201356-7v3q4wn2/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T20:13:56.456581615+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmptlivccoa/port-3976816.txt","pid":3976816,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T20:13:56.456871135+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T20:13:56.458064415+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":3976816} +{"time":"2025-09-16T20:13:56.45807192+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-3976816-3977503-1684149702/socket","Net":"unix"}} +{"time":"2025-09-16T20:13:56.646337913+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T20:13:56.65621458+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"7v3q4wn2","id":"1(@)"} +{"time":"2025-09-16T20:13:57.135083608+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"7v3q4wn2","id":"1(@)"} +{"time":"2025-09-16T20:16:41.311809545+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_201356-7v3q4wn2/logs/debug-internal.log b/wandb/run-20250916_201356-7v3q4wn2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..90f1ac4689a70ff052332ed7d2d5285387ee7049 --- /dev/null +++ b/wandb/run-20250916_201356-7v3q4wn2/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T20:13:56.656394996+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T20:13:56.688512046+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T20:13:57.13496184+08:00","level":"INFO","msg":"stream: created new stream","id":"7v3q4wn2"} +{"time":"2025-09-16T20:13:57.135068844+08:00","level":"INFO","msg":"stream: started","id":"7v3q4wn2"} +{"time":"2025-09-16T20:13:57.135104704+08:00","level":"INFO","msg":"writer: started","stream_id":"7v3q4wn2"} +{"time":"2025-09-16T20:13:57.135160387+08:00","level":"INFO","msg":"handler: started","stream_id":"7v3q4wn2"} +{"time":"2025-09-16T20:13:57.135135274+08:00","level":"INFO","msg":"sender: started","stream_id":"7v3q4wn2"} diff --git a/wandb/run-20250916_201356-7v3q4wn2/logs/debug.log b/wandb/run-20250916_201356-7v3q4wn2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..cd37ec451cb8cff8111997d6ab0be8107b108e36 --- /dev/null +++ b/wandb/run-20250916_201356-7v3q4wn2/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 20:13:56,424 INFO MainThread:3976816 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 20:13:56,424 INFO MainThread:3976816 [wandb_setup.py:_flush():81] Configure stats pid to 3976816 +2025-09-16 20:13:56,424 INFO MainThread:3976816 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_201356-7v3q4wn2/logs/debug.log +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_201356-7v3q4wn2/logs/debug-internal.log +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_init.py:init():813] calling init triggers +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 20:13:56,425 INFO MainThread:3976816 [wandb_init.py:init():854] starting backend +2025-09-16 20:13:56,646 INFO MainThread:3976816 [wandb_init.py:init():857] sending inform_init request +2025-09-16 20:13:56,651 INFO MainThread:3976816 [wandb_init.py:init():865] backend started and connected +2025-09-16 20:13:56,654 INFO MainThread:3976816 [wandb_init.py:init():936] updated telemetry +2025-09-16 20:13:56,666 INFO MainThread:3976816 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 20:13:57,607 INFO MainThread:3976816 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 20:13:57,777 INFO MainThread:3976816 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 20:13:57,777 INFO MainThread:3976816 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 20:13:57,777 INFO MainThread:3976816 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 20:13:57,777 INFO MainThread:3976816 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 20:13:57,780 INFO MainThread:3976816 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 20:15:05,487 INFO MainThread:3976816 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_201356-7v3q4wn2/run-7v3q4wn2.wandb b/wandb/run-20250916_201356-7v3q4wn2/run-7v3q4wn2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_201852-xtyghlyl/files/output.log b/wandb/run-20250916_201852-xtyghlyl/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4c56cea5c5e96cef5c3acdfa5726492bfd07c438 --- /dev/null +++ b/wandb/run-20250916_201852-xtyghlyl/files/output.log @@ -0,0 +1,11 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00@: read: connection reset by peer","id":"1(@)"} +{"time":"2025-09-16T20:35:03.259231056+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_203216-pk1zaawi/logs/debug-internal.log b/wandb/run-20250916_203216-pk1zaawi/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..5f554a719594f319fee76616108d80aef89ab86f --- /dev/null +++ b/wandb/run-20250916_203216-pk1zaawi/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T20:32:16.60690733+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T20:32:16.639477865+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T20:32:17.081092394+08:00","level":"INFO","msg":"stream: created new stream","id":"pk1zaawi"} +{"time":"2025-09-16T20:32:17.081216588+08:00","level":"INFO","msg":"stream: started","id":"pk1zaawi"} +{"time":"2025-09-16T20:32:17.081241424+08:00","level":"INFO","msg":"handler: started","stream_id":"pk1zaawi"} +{"time":"2025-09-16T20:32:17.081251619+08:00","level":"INFO","msg":"sender: started","stream_id":"pk1zaawi"} +{"time":"2025-09-16T20:32:17.081284791+08:00","level":"INFO","msg":"writer: started","stream_id":"pk1zaawi"} diff --git a/wandb/run-20250916_203216-pk1zaawi/logs/debug.log b/wandb/run-20250916_203216-pk1zaawi/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c0fe0af4085747b0e72f7bc3bec9971fb675ef16 --- /dev/null +++ b/wandb/run-20250916_203216-pk1zaawi/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 20:32:16,375 INFO MainThread:4022287 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_setup.py:_flush():81] Configure stats pid to 4022287 +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_203216-pk1zaawi/logs/debug.log +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_203216-pk1zaawi/logs/debug-internal.log +2025-09-16 20:32:16,376 INFO MainThread:4022287 [wandb_init.py:init():813] calling init triggers +2025-09-16 20:32:16,377 INFO MainThread:4022287 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 20:32:16,377 INFO MainThread:4022287 [wandb_init.py:init():854] starting backend +2025-09-16 20:32:16,597 INFO MainThread:4022287 [wandb_init.py:init():857] sending inform_init request +2025-09-16 20:32:16,601 INFO MainThread:4022287 [wandb_init.py:init():865] backend started and connected +2025-09-16 20:32:16,604 INFO MainThread:4022287 [wandb_init.py:init():936] updated telemetry +2025-09-16 20:32:16,615 INFO MainThread:4022287 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 20:32:17,591 INFO MainThread:4022287 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 20:32:17,751 INFO MainThread:4022287 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 20:32:17,753 INFO MainThread:4022287 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 20:32:17,753 INFO MainThread:4022287 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 20:32:17,753 INFO MainThread:4022287 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 20:32:17,756 INFO MainThread:4022287 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 20:33:26,965 INFO MainThread:4022287 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_203216-pk1zaawi/run-pk1zaawi.wandb b/wandb/run-20250916_203216-pk1zaawi/run-pk1zaawi.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_203935-op867fbp/files/config.yaml b/wandb/run-20250916_203935-op867fbp/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..51347d65a424f14d7f65d05e4b9ed8e926c3f3ea --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/files/config.yaml @@ -0,0 +1,91 @@ +_wandb: + value: + cli_version: 0.21.4 + e: + o155rcmkzsw5ssvcc9pikg6mhwlvx0s3: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3576956497920" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T12:39:35.474504Z" + writerId: o155rcmkzsw5ssvcc9pikg6mhwlvx0s3 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 diff --git a/wandb/run-20250916_203935-op867fbp/files/output.log b/wandb/run-20250916_203935-op867fbp/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1d146d64e30941dbeaf769f726d1c54131660c37 --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/files/output.log @@ -0,0 +1,2 @@ + +Detected KeyboardInterrupt, attempting graceful shutdown ... diff --git a/wandb/run-20250916_203935-op867fbp/files/requirements.txt b/wandb/run-20250916_203935-op867fbp/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_203935-op867fbp/files/wandb-metadata.json b/wandb/run-20250916_203935-op867fbp/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3d36088a5466312e21a7284d3f9587b54e73ee2d --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T12:39:35.474504Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576956497920" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "o155rcmkzsw5ssvcc9pikg6mhwlvx0s3" +} \ No newline at end of file diff --git a/wandb/run-20250916_203935-op867fbp/files/wandb-summary.json b/wandb/run-20250916_203935-op867fbp/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..20b4385da45dc73f28a83368c93d1645d8267473 --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":8,"_wandb":{"runtime":8}} \ No newline at end of file diff --git a/wandb/run-20250916_203935-op867fbp/logs/debug-core.log b/wandb/run-20250916_203935-op867fbp/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..22aab800cc89d52b30f7000c5379730ad6eb8be5 --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/logs/debug-core.log @@ -0,0 +1,13 @@ +{"time":"2025-09-16T20:39:35.521624052+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp2zn5m8hn/port-4039164.txt","pid":4039164,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T20:39:35.521932678+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T20:39:35.523628417+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4039164} +{"time":"2025-09-16T20:39:35.523623545+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4039164-4040948-1187046540/socket","Net":"unix"}} +{"time":"2025-09-16T20:39:35.706349004+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T20:39:35.717564512+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"op867fbp","id":"1(@)"} +{"time":"2025-09-16T20:39:36.228432537+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"op867fbp","id":"1(@)"} +{"time":"2025-09-16T20:39:46.018305099+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T20:39:46.018397705+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T20:39:46.018547477+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T20:39:46.018598921+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T20:39:46.019064137+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4039164-4040948-1187046540/socket","Net":"unix"}} +{"time":"2025-09-16T20:39:48.670946886+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_203935-op867fbp/logs/debug-internal.log b/wandb/run-20250916_203935-op867fbp/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..928e9c637c36a122577908471d72321358a4c81f --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/logs/debug-internal.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T20:39:35.717935749+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T20:39:35.753893205+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T20:39:36.228341358+08:00","level":"INFO","msg":"stream: created new stream","id":"op867fbp"} +{"time":"2025-09-16T20:39:36.228416826+08:00","level":"INFO","msg":"stream: started","id":"op867fbp"} +{"time":"2025-09-16T20:39:36.228439132+08:00","level":"INFO","msg":"sender: started","stream_id":"op867fbp"} +{"time":"2025-09-16T20:39:36.228443197+08:00","level":"INFO","msg":"writer: started","stream_id":"op867fbp"} +{"time":"2025-09-16T20:39:36.228490674+08:00","level":"INFO","msg":"handler: started","stream_id":"op867fbp"} +{"time":"2025-09-16T20:39:46.018389019+08:00","level":"INFO","msg":"stream: closing","id":"op867fbp"} diff --git a/wandb/run-20250916_203935-op867fbp/logs/debug.log b/wandb/run-20250916_203935-op867fbp/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7066d461e13b3c661e1266dac2be45e0fe1fd5b1 --- /dev/null +++ b/wandb/run-20250916_203935-op867fbp/logs/debug.log @@ -0,0 +1,23 @@ +2025-09-16 20:39:35,480 INFO MainThread:4039164 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_setup.py:_flush():81] Configure stats pid to 4039164 +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_203935-op867fbp/logs/debug.log +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_203935-op867fbp/logs/debug-internal.log +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_init.py:init():813] calling init triggers +2025-09-16 20:39:35,481 INFO MainThread:4039164 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 20:39:35,482 INFO MainThread:4039164 [wandb_init.py:init():854] starting backend +2025-09-16 20:39:35,706 INFO MainThread:4039164 [wandb_init.py:init():857] sending inform_init request +2025-09-16 20:39:35,711 INFO MainThread:4039164 [wandb_init.py:init():865] backend started and connected +2025-09-16 20:39:35,715 INFO MainThread:4039164 [wandb_init.py:init():936] updated telemetry +2025-09-16 20:39:35,727 INFO MainThread:4039164 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 20:39:37,278 INFO MainThread:4039164 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 20:39:37,447 INFO MainThread:4039164 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 20:39:37,447 INFO MainThread:4039164 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 20:39:37,448 INFO MainThread:4039164 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 20:39:37,448 INFO MainThread:4039164 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 20:39:37,452 INFO MainThread:4039164 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 20:39:46,018 INFO wandb-AsyncioManager-main:4039164 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 20:39:46,018 INFO wandb-AsyncioManager-main:4039164 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_203935-op867fbp/run-op867fbp.wandb b/wandb/run-20250916_203935-op867fbp/run-op867fbp.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_204039-rclulz6c/files/output.log b/wandb/run-20250916_204039-rclulz6c/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c357e8952b1607b0199b0f84df84a082843151b4 --- /dev/null +++ b/wandb/run-20250916_204039-rclulz6c/files/output.log @@ -0,0 +1,11 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 513, in training_step + with autocast(dtype=torch.float16): +NameError: name 'autocast' is not defined +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 513, in training_step +[rank0]: with autocast(dtype=torch.float16): +[rank0]: NameError: name 'autocast' is not defined diff --git a/wandb/run-20250916_212048-tdq91hau/files/requirements.txt b/wandb/run-20250916_212048-tdq91hau/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_212048-tdq91hau/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_212048-tdq91hau/files/wandb-metadata.json b/wandb/run-20250916_212048-tdq91hau/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c0f90f84de7e2b7b38a573fc34d887bb9d4d54ee --- /dev/null +++ b/wandb/run-20250916_212048-tdq91hau/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T13:20:48.657787Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576960593920" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "erd41l5u84rd9czcrih1qpjy9gdrqoqv" +} \ No newline at end of file diff --git a/wandb/run-20250916_212048-tdq91hau/files/wandb-summary.json b/wandb/run-20250916_212048-tdq91hau/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..4f652bc6854b940ff0151f6abf582a7faacfaac8 --- /dev/null +++ b/wandb/run-20250916_212048-tdq91hau/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":69},"_runtime":69} \ No newline at end of file diff --git a/wandb/run-20250916_212048-tdq91hau/logs/debug-core.log b/wandb/run-20250916_212048-tdq91hau/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..d1f644dd2dae45c85dc3422be45e010f600b0e7d --- /dev/null +++ b/wandb/run-20250916_212048-tdq91hau/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T21:20:48.693112378+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmps9itb27b/port-4143575.txt","pid":4143575,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T21:20:48.693323193+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T21:20:48.694517817+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4143575} +{"time":"2025-09-16T21:20:48.694503649+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4143575-4144234-1951090960/socket","Net":"unix"}} +{"time":"2025-09-16T21:20:48.881975988+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T21:20:48.890064023+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"tdq91hau","id":"1(@)"} +{"time":"2025-09-16T21:20:49.381286966+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"tdq91hau","id":"1(@)"} +{"time":"2025-09-16T21:21:58.928562235+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T21:21:58.928629509+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T21:21:58.928678674+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T21:21:58.928693355+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T21:21:58.928850337+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-4143575-4144234-1951090960/socket","Net":"unix"}} +{"time":"2025-09-16T21:22:00.729702477+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T21:22:00.729759411+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T21:22:00.729790979+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_212048-tdq91hau/logs/debug-internal.log b/wandb/run-20250916_212048-tdq91hau/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2e77f7c92541e875fa6fa4f27fedeeaa76b52b95 --- /dev/null +++ b/wandb/run-20250916_212048-tdq91hau/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T21:20:48.890224649+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T21:20:48.915264294+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T21:20:49.381140263+08:00","level":"INFO","msg":"stream: created new stream","id":"tdq91hau"} +{"time":"2025-09-16T21:20:49.381272591+08:00","level":"INFO","msg":"stream: started","id":"tdq91hau"} +{"time":"2025-09-16T21:20:49.381299056+08:00","level":"INFO","msg":"handler: started","stream_id":"tdq91hau"} +{"time":"2025-09-16T21:20:49.381300624+08:00","level":"INFO","msg":"writer: started","stream_id":"tdq91hau"} +{"time":"2025-09-16T21:20:49.381322908+08:00","level":"INFO","msg":"sender: started","stream_id":"tdq91hau"} +{"time":"2025-09-16T21:21:58.928636265+08:00","level":"INFO","msg":"stream: closing","id":"tdq91hau"} +{"time":"2025-09-16T21:22:00.239779352+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T21:22:00.727857736+08:00","level":"INFO","msg":"handler: closed","stream_id":"tdq91hau"} +{"time":"2025-09-16T21:22:00.728462376+08:00","level":"INFO","msg":"sender: closed","stream_id":"tdq91hau"} +{"time":"2025-09-16T21:22:00.728515986+08:00","level":"INFO","msg":"stream: closed","id":"tdq91hau"} diff --git a/wandb/run-20250916_212048-tdq91hau/logs/debug.log b/wandb/run-20250916_212048-tdq91hau/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..abe6b2a4791db4614bdf5eb57add7b698c333025 --- /dev/null +++ b/wandb/run-20250916_212048-tdq91hau/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 21:20:48,661 INFO MainThread:4143575 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 21:20:48,661 INFO MainThread:4143575 [wandb_setup.py:_flush():81] Configure stats pid to 4143575 +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_212048-tdq91hau/logs/debug.log +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_212048-tdq91hau/logs/debug-internal.log +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_init.py:init():813] calling init triggers +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 21:20:48,662 INFO MainThread:4143575 [wandb_init.py:init():854] starting backend +2025-09-16 21:20:48,882 INFO MainThread:4143575 [wandb_init.py:init():857] sending inform_init request +2025-09-16 21:20:48,886 INFO MainThread:4143575 [wandb_init.py:init():865] backend started and connected +2025-09-16 21:20:48,889 INFO MainThread:4143575 [wandb_init.py:init():936] updated telemetry +2025-09-16 21:20:48,900 INFO MainThread:4143575 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 21:20:49,704 INFO MainThread:4143575 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 21:20:49,873 INFO MainThread:4143575 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 21:20:49,873 INFO MainThread:4143575 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 21:20:49,873 INFO MainThread:4143575 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 21:20:49,873 INFO MainThread:4143575 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 21:20:49,875 INFO MainThread:4143575 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 21:21:58,396 INFO MainThread:4143575 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 21:21:58,928 INFO wandb-AsyncioManager-main:4143575 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 21:21:58,928 INFO wandb-AsyncioManager-main:4143575 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_212048-tdq91hau/run-tdq91hau.wandb b/wandb/run-20250916_212048-tdq91hau/run-tdq91hau.wandb new file mode 100644 index 0000000000000000000000000000000000000000..6b23bedcaab741d4e9ffdb9936e035c76d650211 Binary files /dev/null and b/wandb/run-20250916_212048-tdq91hau/run-tdq91hau.wandb differ diff --git a/wandb/run-20250916_212323-9uaee180/files/output.log b/wandb/run-20250916_212323-9uaee180/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4193b122c6621dd72286a97bafcf029dd6ab28a0 --- /dev/null +++ b/wandb/run-20250916_212323-9uaee180/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:51<11:47:48, 0.04it/s, v_num=e180] diff --git a/wandb/run-20250916_212323-9uaee180/files/requirements.txt b/wandb/run-20250916_212323-9uaee180/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_212323-9uaee180/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_212323-9uaee180/files/wandb-metadata.json b/wandb/run-20250916_212323-9uaee180/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..86c0ccc86fe5a80e5e07703f18e2269b892fa2ac --- /dev/null +++ b/wandb/run-20250916_212323-9uaee180/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T13:23:23.991735Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576961064960" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "sup1rea96ue9hnwrju294kyplylyqwms" +} \ No newline at end of file diff --git a/wandb/run-20250916_212323-9uaee180/logs/debug-core.log b/wandb/run-20250916_212323-9uaee180/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..683362e0441a45a52bd9afa457a7c29d7b317af9 --- /dev/null +++ b/wandb/run-20250916_212323-9uaee180/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T21:23:24.027328236+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp743wkhb3/port-4150181.txt","pid":4150181,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T21:23:24.027600633+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T21:23:24.028323735+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4150181} +{"time":"2025-09-16T21:23:24.02831835+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4150181-4150863-4217112134/socket","Net":"unix"}} +{"time":"2025-09-16T21:23:24.216595974+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T21:23:24.228699469+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"9uaee180","id":"1(@)"} +{"time":"2025-09-16T21:23:24.705658771+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"9uaee180","id":"1(@)"} +{"time":"2025-09-16T21:26:09.365728675+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_212323-9uaee180/logs/debug-internal.log b/wandb/run-20250916_212323-9uaee180/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..3f63ac72f449ffbee839a25ea7477a7cae87147d --- /dev/null +++ b/wandb/run-20250916_212323-9uaee180/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T21:23:24.228882899+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T21:23:24.244930214+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T21:23:24.705544663+08:00","level":"INFO","msg":"stream: created new stream","id":"9uaee180"} +{"time":"2025-09-16T21:23:24.705642948+08:00","level":"INFO","msg":"stream: started","id":"9uaee180"} +{"time":"2025-09-16T21:23:24.705683616+08:00","level":"INFO","msg":"writer: started","stream_id":"9uaee180"} +{"time":"2025-09-16T21:23:24.705789793+08:00","level":"INFO","msg":"handler: started","stream_id":"9uaee180"} +{"time":"2025-09-16T21:23:24.705839992+08:00","level":"INFO","msg":"sender: started","stream_id":"9uaee180"} diff --git a/wandb/run-20250916_212323-9uaee180/logs/debug.log b/wandb/run-20250916_212323-9uaee180/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..19410f73819819a5a6482eb4b58968ff45cd563f --- /dev/null +++ b/wandb/run-20250916_212323-9uaee180/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 21:23:23,995 INFO MainThread:4150181 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_setup.py:_flush():81] Configure stats pid to 4150181 +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_212323-9uaee180/logs/debug.log +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_212323-9uaee180/logs/debug-internal.log +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_init.py:init():813] calling init triggers +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 21:23:23,996 INFO MainThread:4150181 [wandb_init.py:init():854] starting backend +2025-09-16 21:23:24,216 INFO MainThread:4150181 [wandb_init.py:init():857] sending inform_init request +2025-09-16 21:23:24,221 INFO MainThread:4150181 [wandb_init.py:init():865] backend started and connected +2025-09-16 21:23:24,224 INFO MainThread:4150181 [wandb_init.py:init():936] updated telemetry +2025-09-16 21:23:24,235 INFO MainThread:4150181 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 21:23:24,994 INFO MainThread:4150181 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 21:23:25,180 INFO MainThread:4150181 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 21:23:25,180 INFO MainThread:4150181 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 21:23:25,180 INFO MainThread:4150181 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 21:23:25,180 INFO MainThread:4150181 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 21:23:25,183 INFO MainThread:4150181 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 21:24:34,725 INFO MainThread:4150181 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_212323-9uaee180/run-9uaee180.wandb b/wandb/run-20250916_212323-9uaee180/run-9uaee180.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_212834-zqm3zk0g/files/output.log b/wandb/run-20250916_212834-zqm3zk0g/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ce8dac5223a28124e4e05ad40e5bac4502ba8b65 --- /dev/null +++ b/wandb/run-20250916_212834-zqm3zk0g/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:55<12:48:56, 0.04it/s, v_num=zk0g] diff --git a/wandb/run-20250916_212834-zqm3zk0g/files/requirements.txt b/wandb/run-20250916_212834-zqm3zk0g/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_212834-zqm3zk0g/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_212834-zqm3zk0g/files/wandb-metadata.json b/wandb/run-20250916_212834-zqm3zk0g/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c2fe5448bd0ee6476e05960c756c89f90c8045fa --- /dev/null +++ b/wandb/run-20250916_212834-zqm3zk0g/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T13:28:34.462200Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576961589248" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "pkij8iz2xyrvs10w3syao8upltrviba5" +} \ No newline at end of file diff --git a/wandb/run-20250916_212834-zqm3zk0g/logs/debug-core.log b/wandb/run-20250916_212834-zqm3zk0g/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..41d7f1f3e936b271ca8405079e0a4c0d58afb3b0 --- /dev/null +++ b/wandb/run-20250916_212834-zqm3zk0g/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T21:28:34.496774025+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpk4i0zchd/port-4163915.txt","pid":4163915,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T21:28:34.49698474+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T21:28:34.497507496+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4163915} +{"time":"2025-09-16T21:28:34.49750023+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4163915-4165110-3937545774/socket","Net":"unix"}} +{"time":"2025-09-16T21:28:34.685663325+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T21:28:34.692538455+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"zqm3zk0g","id":"1(@)"} +{"time":"2025-09-16T21:28:35.140318846+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"zqm3zk0g","id":"1(@)"} +{"time":"2025-09-16T21:31:26.787323344+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_212834-zqm3zk0g/logs/debug-internal.log b/wandb/run-20250916_212834-zqm3zk0g/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..79f5a9176c4b8c5f7cbcf89390cac2836e9a47a9 --- /dev/null +++ b/wandb/run-20250916_212834-zqm3zk0g/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T21:28:34.692657129+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T21:28:34.707416279+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T21:28:35.140156942+08:00","level":"INFO","msg":"stream: created new stream","id":"zqm3zk0g"} +{"time":"2025-09-16T21:28:35.14029648+08:00","level":"INFO","msg":"stream: started","id":"zqm3zk0g"} +{"time":"2025-09-16T21:28:35.14033938+08:00","level":"INFO","msg":"writer: started","stream_id":"zqm3zk0g"} +{"time":"2025-09-16T21:28:35.14036248+08:00","level":"INFO","msg":"sender: started","stream_id":"zqm3zk0g"} +{"time":"2025-09-16T21:28:35.140411401+08:00","level":"INFO","msg":"handler: started","stream_id":"zqm3zk0g"} diff --git a/wandb/run-20250916_212834-zqm3zk0g/logs/debug.log b/wandb/run-20250916_212834-zqm3zk0g/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..2a2ad2fbaf483344c4f35957e4bd9c60a65d86af --- /dev/null +++ b/wandb/run-20250916_212834-zqm3zk0g/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_setup.py:_flush():81] Configure stats pid to 4163915 +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_212834-zqm3zk0g/logs/debug.log +2025-09-16 21:28:34,466 INFO MainThread:4163915 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_212834-zqm3zk0g/logs/debug-internal.log +2025-09-16 21:28:34,467 INFO MainThread:4163915 [wandb_init.py:init():813] calling init triggers +2025-09-16 21:28:34,467 INFO MainThread:4163915 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 21:28:34,467 INFO MainThread:4163915 [wandb_init.py:init():854] starting backend +2025-09-16 21:28:34,685 INFO MainThread:4163915 [wandb_init.py:init():857] sending inform_init request +2025-09-16 21:28:34,690 INFO MainThread:4163915 [wandb_init.py:init():865] backend started and connected +2025-09-16 21:28:34,693 INFO MainThread:4163915 [wandb_init.py:init():936] updated telemetry +2025-09-16 21:28:34,703 INFO MainThread:4163915 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 21:28:35,451 INFO MainThread:4163915 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 21:28:35,613 INFO MainThread:4163915 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 21:28:35,614 INFO MainThread:4163915 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 21:28:35,614 INFO MainThread:4163915 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 21:28:35,614 INFO MainThread:4163915 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 21:28:35,617 INFO MainThread:4163915 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 21:29:51,342 INFO MainThread:4163915 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_212834-zqm3zk0g/run-zqm3zk0g.wandb b/wandb/run-20250916_212834-zqm3zk0g/run-zqm3zk0g.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_213218-ggiwlop2/files/output.log b/wandb/run-20250916_213218-ggiwlop2/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..44d6b30724cef3ff20862aec58c22518b68ac17b --- /dev/null +++ b/wandb/run-20250916_213218-ggiwlop2/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:51<11:54:57, 0.04it/s, v_num=lop2] diff --git a/wandb/run-20250916_213218-ggiwlop2/files/requirements.txt b/wandb/run-20250916_213218-ggiwlop2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_213218-ggiwlop2/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_213218-ggiwlop2/files/wandb-metadata.json b/wandb/run-20250916_213218-ggiwlop2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8f675d0b097b86610a9261d4c62adbe94b29db37 --- /dev/null +++ b/wandb/run-20250916_213218-ggiwlop2/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T13:32:18.163742Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576962035712" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "vm6aaurjkdrnwguz5cszawnq348uobr1" +} \ No newline at end of file diff --git a/wandb/run-20250916_213218-ggiwlop2/logs/debug-core.log b/wandb/run-20250916_213218-ggiwlop2/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..67666c88fddee520379a2174632f70a58f9678de --- /dev/null +++ b/wandb/run-20250916_213218-ggiwlop2/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T21:32:18.210157128+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpr51e7fq6/port-4174514.txt","pid":4174514,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T21:32:18.210462163+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T21:32:18.211484074+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4174514-4175206-561920732/socket","Net":"unix"}} +{"time":"2025-09-16T21:32:18.211628155+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4174514} +{"time":"2025-09-16T21:32:18.393651755+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T21:32:18.409975803+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"ggiwlop2","id":"1(@)"} +{"time":"2025-09-16T21:32:18.907675905+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"ggiwlop2","id":"1(@)"} +{"time":"2025-09-16T21:35:04.071843783+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_213218-ggiwlop2/logs/debug-internal.log b/wandb/run-20250916_213218-ggiwlop2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..ad8da4344311cdcb19e7fe608f391f692b96ec2e --- /dev/null +++ b/wandb/run-20250916_213218-ggiwlop2/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T21:32:18.410182359+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T21:32:18.437924069+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T21:32:18.907489404+08:00","level":"INFO","msg":"stream: created new stream","id":"ggiwlop2"} +{"time":"2025-09-16T21:32:18.907659458+08:00","level":"INFO","msg":"stream: started","id":"ggiwlop2"} +{"time":"2025-09-16T21:32:18.907698345+08:00","level":"INFO","msg":"writer: started","stream_id":"ggiwlop2"} +{"time":"2025-09-16T21:32:18.907929612+08:00","level":"INFO","msg":"handler: started","stream_id":"ggiwlop2"} +{"time":"2025-09-16T21:32:18.908071186+08:00","level":"INFO","msg":"sender: started","stream_id":"ggiwlop2"} diff --git a/wandb/run-20250916_213218-ggiwlop2/logs/debug.log b/wandb/run-20250916_213218-ggiwlop2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..092968992a9923d3d89c252b429547d2a833c8be --- /dev/null +++ b/wandb/run-20250916_213218-ggiwlop2/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 21:32:18,169 INFO MainThread:4174514 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 21:32:18,169 INFO MainThread:4174514 [wandb_setup.py:_flush():81] Configure stats pid to 4174514 +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_213218-ggiwlop2/logs/debug.log +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_213218-ggiwlop2/logs/debug-internal.log +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_init.py:init():813] calling init triggers +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 21:32:18,170 INFO MainThread:4174514 [wandb_init.py:init():854] starting backend +2025-09-16 21:32:18,394 INFO MainThread:4174514 [wandb_init.py:init():857] sending inform_init request +2025-09-16 21:32:18,400 INFO MainThread:4174514 [wandb_init.py:init():865] backend started and connected +2025-09-16 21:32:18,405 INFO MainThread:4174514 [wandb_init.py:init():936] updated telemetry +2025-09-16 21:32:18,418 INFO MainThread:4174514 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 21:32:19,331 INFO MainThread:4174514 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 21:32:19,515 INFO MainThread:4174514 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 21:32:19,515 INFO MainThread:4174514 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 21:32:19,516 INFO MainThread:4174514 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 21:32:19,516 INFO MainThread:4174514 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 21:32:19,519 INFO MainThread:4174514 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 21:33:28,560 INFO MainThread:4174514 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_213218-ggiwlop2/run-ggiwlop2.wandb b/wandb/run-20250916_213218-ggiwlop2/run-ggiwlop2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_214214-qdk5d6zx/files/output.log b/wandb/run-20250916_214214-qdk5d6zx/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..408327b68640bdf2e02abbfe47ad14fe9fb794c5 --- /dev/null +++ b/wandb/run-20250916_214214-qdk5d6zx/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:52<12:01:12, 0.04it/s, v_num=d6zx] diff --git a/wandb/run-20250916_214214-qdk5d6zx/files/requirements.txt b/wandb/run-20250916_214214-qdk5d6zx/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_214214-qdk5d6zx/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_214214-qdk5d6zx/files/wandb-metadata.json b/wandb/run-20250916_214214-qdk5d6zx/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8e10e369108b717478ab4ab6d6ab5809298eba1d --- /dev/null +++ b/wandb/run-20250916_214214-qdk5d6zx/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T13:42:14.588798Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576962539520" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "97zmh6pl6v8jcfkbjb41cwum05qyhsc7" +} \ No newline at end of file diff --git a/wandb/run-20250916_214214-qdk5d6zx/logs/debug-core.log b/wandb/run-20250916_214214-qdk5d6zx/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..5a7ce906ea79d5f97222caf4cccd93f201350069 --- /dev/null +++ b/wandb/run-20250916_214214-qdk5d6zx/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T21:42:14.624257006+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpq00b4q4i/port-4294.txt","pid":4294,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T21:42:14.624520754+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T21:42:14.626575532+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-4294-5246-3340261787/socket","Net":"unix"}} +{"time":"2025-09-16T21:42:14.626667487+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":4294} +{"time":"2025-09-16T21:42:14.813573587+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T21:42:14.822532134+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"qdk5d6zx","id":"1(@)"} +{"time":"2025-09-16T21:42:15.285085975+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"qdk5d6zx","id":"1(@)"} +{"time":"2025-09-16T21:45:00.959817478+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_214214-qdk5d6zx/logs/debug-internal.log b/wandb/run-20250916_214214-qdk5d6zx/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..fefde16be5074585ba009672f9676ac89f072ae7 --- /dev/null +++ b/wandb/run-20250916_214214-qdk5d6zx/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T21:42:14.822693676+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T21:42:14.849544951+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T21:42:15.284969282+08:00","level":"INFO","msg":"stream: created new stream","id":"qdk5d6zx"} +{"time":"2025-09-16T21:42:15.28506871+08:00","level":"INFO","msg":"stream: started","id":"qdk5d6zx"} +{"time":"2025-09-16T21:42:15.285103997+08:00","level":"INFO","msg":"writer: started","stream_id":"qdk5d6zx"} +{"time":"2025-09-16T21:42:15.285144803+08:00","level":"INFO","msg":"handler: started","stream_id":"qdk5d6zx"} +{"time":"2025-09-16T21:42:15.285115504+08:00","level":"INFO","msg":"sender: started","stream_id":"qdk5d6zx"} diff --git a/wandb/run-20250916_214214-qdk5d6zx/logs/debug.log b/wandb/run-20250916_214214-qdk5d6zx/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6c63a305342d38037191be742b7c376f303554d4 --- /dev/null +++ b/wandb/run-20250916_214214-qdk5d6zx/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 21:42:14,592 INFO MainThread:4294 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 21:42:14,592 INFO MainThread:4294 [wandb_setup.py:_flush():81] Configure stats pid to 4294 +2025-09-16 21:42:14,593 INFO MainThread:4294 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 21:42:14,593 INFO MainThread:4294 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 21:42:14,593 INFO MainThread:4294 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 21:42:14,593 INFO MainThread:4294 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_214214-qdk5d6zx/logs/debug.log +2025-09-16 21:42:14,593 INFO MainThread:4294 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_214214-qdk5d6zx/logs/debug-internal.log +2025-09-16 21:42:14,593 INFO MainThread:4294 [wandb_init.py:init():813] calling init triggers +2025-09-16 21:42:14,594 INFO MainThread:4294 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 21:42:14,594 INFO MainThread:4294 [wandb_init.py:init():854] starting backend +2025-09-16 21:42:14,813 INFO MainThread:4294 [wandb_init.py:init():857] sending inform_init request +2025-09-16 21:42:14,818 INFO MainThread:4294 [wandb_init.py:init():865] backend started and connected +2025-09-16 21:42:14,821 INFO MainThread:4294 [wandb_init.py:init():936] updated telemetry +2025-09-16 21:42:14,831 INFO MainThread:4294 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 21:42:15,648 INFO MainThread:4294 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 21:42:15,814 INFO MainThread:4294 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 21:42:15,814 INFO MainThread:4294 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 21:42:15,814 INFO MainThread:4294 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 21:42:15,814 INFO MainThread:4294 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 21:42:15,816 INFO MainThread:4294 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 21:43:24,877 INFO MainThread:4294 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_214214-qdk5d6zx/run-qdk5d6zx.wandb b/wandb/run-20250916_214214-qdk5d6zx/run-qdk5d6zx.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_215226-bkyyj1s3/files/output.log b/wandb/run-20250916_215226-bkyyj1s3/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..2ba2f234cb549a514e33ce7454eb5ff38d640e5b --- /dev/null +++ b/wandb/run-20250916_215226-bkyyj1s3/files/output.log @@ -0,0 +1,13 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00 + cli_main() + File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main + cli = LightningCLI( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ + self._run_subcommand(self.subcommand) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand + fn(**fn_kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit + call._call_and_handle_interrupt( + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt + return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch + return function(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl + self._run(model, ckpt_path=ckpt_path) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run + results = self._run_stage() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage + self.fit_loop.run() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run + self.advance() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance + self.epoch_loop.run(self._data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run + self.advance(data_fetcher) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance + batch_output = self.manual_optimization.run(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run + self.advance(kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance + training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook + output = fn(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step + return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ + wrapper_output = wrapper_module(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl + return forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn + ret_val = func(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward + loss = self.module(*inputs, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl + return self._call_impl(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl + return inner() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner + result = forward_call(*args, **kwargs) + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward + out = method(*_args, **_kwargs) + File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 537, in training_step + critic_state = self.critic.model.state_dict() + File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ + raise AttributeError( +AttributeError: 'RobertaCritic' object has no attribute 'model' +[rank0]: Traceback (most recent call last): +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 33, in +[rank0]: cli_main() +[rank0]: File "/home/jiashuo/codes/OfflineArcher/main.py", line 22, in cli_main +[rank0]: cli = LightningCLI( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 414, in __init__ +[rank0]: self._run_subcommand(self.subcommand) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/cli.py", line 747, in _run_subcommand +[rank0]: fn(**fn_kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 560, in fit +[rank0]: call._call_and_handle_interrupt( +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 48, in _call_and_handle_interrupt +[rank0]: return trainer.strategy.launcher.launch(trainer_fn, *args, trainer=trainer, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/launchers/subprocess_script.py", line 105, in launch +[rank0]: return function(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 598, in _fit_impl +[rank0]: self._run(model, ckpt_path=ckpt_path) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1011, in _run +[rank0]: results = self._run_stage() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/trainer.py", line 1055, in _run_stage +[rank0]: self.fit_loop.run() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 216, in run +[rank0]: self.advance() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py", line 458, in advance +[rank0]: self.epoch_loop.run(self._data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 152, in run +[rank0]: self.advance(data_fetcher) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/training_epoch_loop.py", line 350, in advance +[rank0]: batch_output = self.manual_optimization.run(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 95, in run +[rank0]: self.advance(kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/optimization/manual.py", line 115, in advance +[rank0]: training_step_output = call._call_strategy_hook(trainer, "training_step", *kwargs.values()) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/trainer/call.py", line 329, in _call_strategy_hook +[rank0]: output = fn(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 390, in training_step +[rank0]: return self._forward_redirection(self.model, self.lightning_module, "training_step", *args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 641, in __call__ +[rank0]: wrapper_output = wrapper_module(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1784, in _call_impl +[rank0]: return forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/utils/nvtx.py", line 20, in wrapped_fn +[rank0]: ret_val = func(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/deepspeed/runtime/engine.py", line 2131, in forward +[rank0]: loss = self.module(*inputs, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1773, in _wrapped_call_impl +[rank0]: return self._call_impl(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1879, in _call_impl +[rank0]: return inner() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1827, in inner +[rank0]: result = forward_call(*args, **kwargs) +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/strategies/strategy.py", line 634, in wrapped_forward +[rank0]: out = method(*_args, **_kwargs) +[rank0]: File "/home/jiashuo/codes/OfflineArcher/Algorithms.py", line 537, in training_step +[rank0]: critic_state = self.critic.model.state_dict() +[rank0]: File "/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1962, in __getattr__ +[rank0]: raise AttributeError( +[rank0]: AttributeError: 'RobertaCritic' object has no attribute 'model' diff --git a/wandb/run-20250916_221735-sc93vfxc/files/requirements.txt b/wandb/run-20250916_221735-sc93vfxc/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_221735-sc93vfxc/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_221735-sc93vfxc/files/wandb-metadata.json b/wandb/run-20250916_221735-sc93vfxc/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..38383b32f7490cb799ba02fda0c6e40f36c254dc --- /dev/null +++ b/wandb/run-20250916_221735-sc93vfxc/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T14:17:35.429424Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576965910528" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "adjby25t1hw4o9aoql82pp3l0mnxwaz3" +} \ No newline at end of file diff --git a/wandb/run-20250916_221735-sc93vfxc/files/wandb-summary.json b/wandb/run-20250916_221735-sc93vfxc/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..7db84d93f965f3c6dc610975e4341d4dc6d6319a --- /dev/null +++ b/wandb/run-20250916_221735-sc93vfxc/files/wandb-summary.json @@ -0,0 +1 @@ +{"_runtime":69,"_wandb":{"runtime":69}} \ No newline at end of file diff --git a/wandb/run-20250916_221735-sc93vfxc/logs/debug-core.log b/wandb/run-20250916_221735-sc93vfxc/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..df5e4b65316427d2dc38ef36ae37f5ca0b19837f --- /dev/null +++ b/wandb/run-20250916_221735-sc93vfxc/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-09-16T22:17:35.464840816+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpr8mklxhg/port-85495.txt","pid":85495,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T22:17:35.465051588+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T22:17:35.465621148+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":85495} +{"time":"2025-09-16T22:17:35.465624084+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-85495-86623-621912914/socket","Net":"unix"}} +{"time":"2025-09-16T22:17:35.653545767+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T22:17:35.662131561+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"sc93vfxc","id":"1(@)"} +{"time":"2025-09-16T22:17:36.152887266+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"sc93vfxc","id":"1(@)"} +{"time":"2025-09-16T22:18:46.069260283+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"1(@)"} +{"time":"2025-09-16T22:18:46.069354887+08:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-09-16T22:18:46.069359609+08:00","level":"INFO","msg":"connection: closing","id":"1(@)"} +{"time":"2025-09-16T22:18:46.069496775+08:00","level":"INFO","msg":"connection: closed successfully","id":"1(@)"} +{"time":"2025-09-16T22:18:46.069535673+08:00","level":"INFO","msg":"server: listener closed","addr":{"Name":"/tmp/wandb-85495-86623-621912914/socket","Net":"unix"}} +{"time":"2025-09-16T22:18:47.30041033+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"1(@)"} +{"time":"2025-09-16T22:18:47.300458119+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"1(@)"} +{"time":"2025-09-16T22:18:47.300479638+08:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250916_221735-sc93vfxc/logs/debug-internal.log b/wandb/run-20250916_221735-sc93vfxc/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..3c8993dcf3bd50d9f8e0886d7d96d90e1425aedd --- /dev/null +++ b/wandb/run-20250916_221735-sc93vfxc/logs/debug-internal.log @@ -0,0 +1,12 @@ +{"time":"2025-09-16T22:17:35.662328325+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T22:17:35.687544739+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T22:17:36.152746082+08:00","level":"INFO","msg":"stream: created new stream","id":"sc93vfxc"} +{"time":"2025-09-16T22:17:36.152871041+08:00","level":"INFO","msg":"stream: started","id":"sc93vfxc"} +{"time":"2025-09-16T22:17:36.153027533+08:00","level":"INFO","msg":"sender: started","stream_id":"sc93vfxc"} +{"time":"2025-09-16T22:17:36.153004623+08:00","level":"INFO","msg":"writer: started","stream_id":"sc93vfxc"} +{"time":"2025-09-16T22:17:36.153070233+08:00","level":"INFO","msg":"handler: started","stream_id":"sc93vfxc"} +{"time":"2025-09-16T22:18:46.069376501+08:00","level":"INFO","msg":"stream: closing","id":"sc93vfxc"} +{"time":"2025-09-16T22:18:47.022456229+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-09-16T22:18:47.298108094+08:00","level":"INFO","msg":"handler: closed","stream_id":"sc93vfxc"} +{"time":"2025-09-16T22:18:47.298890728+08:00","level":"INFO","msg":"sender: closed","stream_id":"sc93vfxc"} +{"time":"2025-09-16T22:18:47.298949881+08:00","level":"INFO","msg":"stream: closed","id":"sc93vfxc"} diff --git a/wandb/run-20250916_221735-sc93vfxc/logs/debug.log b/wandb/run-20250916_221735-sc93vfxc/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4399d99b4d535eb93088724a516b90300909b38a --- /dev/null +++ b/wandb/run-20250916_221735-sc93vfxc/logs/debug.log @@ -0,0 +1,24 @@ +2025-09-16 22:17:35,433 INFO MainThread:85495 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 22:17:35,433 INFO MainThread:85495 [wandb_setup.py:_flush():81] Configure stats pid to 85495 +2025-09-16 22:17:35,433 INFO MainThread:85495 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 22:17:35,433 INFO MainThread:85495 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 22:17:35,433 INFO MainThread:85495 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 22:17:35,434 INFO MainThread:85495 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_221735-sc93vfxc/logs/debug.log +2025-09-16 22:17:35,434 INFO MainThread:85495 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_221735-sc93vfxc/logs/debug-internal.log +2025-09-16 22:17:35,434 INFO MainThread:85495 [wandb_init.py:init():813] calling init triggers +2025-09-16 22:17:35,434 INFO MainThread:85495 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 22:17:35,434 INFO MainThread:85495 [wandb_init.py:init():854] starting backend +2025-09-16 22:17:35,653 INFO MainThread:85495 [wandb_init.py:init():857] sending inform_init request +2025-09-16 22:17:35,658 INFO MainThread:85495 [wandb_init.py:init():865] backend started and connected +2025-09-16 22:17:35,661 INFO MainThread:85495 [wandb_init.py:init():936] updated telemetry +2025-09-16 22:17:35,672 INFO MainThread:85495 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 22:17:36,490 INFO MainThread:85495 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 22:17:36,658 INFO MainThread:85495 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 22:17:36,658 INFO MainThread:85495 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 22:17:36,658 INFO MainThread:85495 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 22:17:36,658 INFO MainThread:85495 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 22:17:36,661 INFO MainThread:85495 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 22:18:45,593 INFO MainThread:85495 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} +2025-09-16 22:18:46,069 INFO wandb-AsyncioManager-main:85495 [service_client.py:_forward_responses():84] Reached EOF. +2025-09-16 22:18:46,069 INFO wandb-AsyncioManager-main:85495 [mailbox.py:close():137] Closing mailbox, abandoning 1 handles. diff --git a/wandb/run-20250916_221735-sc93vfxc/run-sc93vfxc.wandb b/wandb/run-20250916_221735-sc93vfxc/run-sc93vfxc.wandb new file mode 100644 index 0000000000000000000000000000000000000000..643dbccf71d5bf3fab04fa372dcd294189c8aa99 Binary files /dev/null and b/wandb/run-20250916_221735-sc93vfxc/run-sc93vfxc.wandb differ diff --git a/wandb/run-20250916_222352-u4st2js4/files/output.log b/wandb/run-20250916_222352-u4st2js4/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..ad3892193f3aa929961924870d6adac52c7b1413 --- /dev/null +++ b/wandb/run-20250916_222352-u4st2js4/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:51<11:54:54, 0.04it/s, v_num=2js4] diff --git a/wandb/run-20250916_222352-u4st2js4/files/requirements.txt b/wandb/run-20250916_222352-u4st2js4/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_222352-u4st2js4/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_222352-u4st2js4/files/wandb-metadata.json b/wandb/run-20250916_222352-u4st2js4/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bb6735dae925d494603b8faeedddac2e31388afb --- /dev/null +++ b/wandb/run-20250916_222352-u4st2js4/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T14:23:52.709581Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576956116992" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "8yjatipasjuihbk7tdf8zudm6ary6cm3" +} \ No newline at end of file diff --git a/wandb/run-20250916_222352-u4st2js4/logs/debug-core.log b/wandb/run-20250916_222352-u4st2js4/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..cfebdb551fa077767a9b35902e7ca04d51cbbf68 --- /dev/null +++ b/wandb/run-20250916_222352-u4st2js4/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T22:23:52.749027208+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpehvchgal/port-101089.txt","pid":101089,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T22:23:52.749278708+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T22:23:52.749854069+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":101089} +{"time":"2025-09-16T22:23:52.749825323+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-101089-101742-3679494040/socket","Net":"unix"}} +{"time":"2025-09-16T22:23:52.938013682+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T22:23:52.947324926+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"u4st2js4","id":"1(@)"} +{"time":"2025-09-16T22:23:53.46957394+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"u4st2js4","id":"1(@)"} +{"time":"2025-09-16T22:26:38.513436465+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_222352-u4st2js4/logs/debug-internal.log b/wandb/run-20250916_222352-u4st2js4/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0830f2cc8427c99c1ecff880e67048da5ca3e32a --- /dev/null +++ b/wandb/run-20250916_222352-u4st2js4/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T22:23:52.947625816+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T22:23:52.97950315+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T22:23:53.469440971+08:00","level":"INFO","msg":"stream: created new stream","id":"u4st2js4"} +{"time":"2025-09-16T22:23:53.469556983+08:00","level":"INFO","msg":"stream: started","id":"u4st2js4"} +{"time":"2025-09-16T22:23:53.469700452+08:00","level":"INFO","msg":"sender: started","stream_id":"u4st2js4"} +{"time":"2025-09-16T22:23:53.469701149+08:00","level":"INFO","msg":"writer: started","stream_id":"u4st2js4"} +{"time":"2025-09-16T22:23:53.469756821+08:00","level":"INFO","msg":"handler: started","stream_id":"u4st2js4"} diff --git a/wandb/run-20250916_222352-u4st2js4/logs/debug.log b/wandb/run-20250916_222352-u4st2js4/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c21d3888c274420b925e92292dbedc48faeadd4a --- /dev/null +++ b/wandb/run-20250916_222352-u4st2js4/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 22:23:52,713 INFO MainThread:101089 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 22:23:52,713 INFO MainThread:101089 [wandb_setup.py:_flush():81] Configure stats pid to 101089 +2025-09-16 22:23:52,713 INFO MainThread:101089 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 22:23:52,713 INFO MainThread:101089 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 22:23:52,713 INFO MainThread:101089 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 22:23:52,714 INFO MainThread:101089 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_222352-u4st2js4/logs/debug.log +2025-09-16 22:23:52,714 INFO MainThread:101089 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_222352-u4st2js4/logs/debug-internal.log +2025-09-16 22:23:52,714 INFO MainThread:101089 [wandb_init.py:init():813] calling init triggers +2025-09-16 22:23:52,714 INFO MainThread:101089 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 22:23:52,714 INFO MainThread:101089 [wandb_init.py:init():854] starting backend +2025-09-16 22:23:52,938 INFO MainThread:101089 [wandb_init.py:init():857] sending inform_init request +2025-09-16 22:23:52,942 INFO MainThread:101089 [wandb_init.py:init():865] backend started and connected +2025-09-16 22:23:52,945 INFO MainThread:101089 [wandb_init.py:init():936] updated telemetry +2025-09-16 22:23:52,956 INFO MainThread:101089 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 22:23:53,855 INFO MainThread:101089 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 22:23:54,022 INFO MainThread:101089 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 22:23:54,023 INFO MainThread:101089 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 22:23:54,024 INFO MainThread:101089 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 22:23:54,024 INFO MainThread:101089 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 22:23:54,028 INFO MainThread:101089 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 22:25:02,871 INFO MainThread:101089 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_222352-u4st2js4/run-u4st2js4.wandb b/wandb/run-20250916_222352-u4st2js4/run-u4st2js4.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_222834-l5x14zzr/files/output.log b/wandb/run-20250916_222834-l5x14zzr/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..3a3cf158610399b5cd2ca430ee2ee8432c6101f3 --- /dev/null +++ b/wandb/run-20250916_222834-l5x14zzr/files/output.log @@ -0,0 +1,9 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 2/1652 [00:52<11:56:45, 0.04it/s, v_num=4zzr] diff --git a/wandb/run-20250916_222834-l5x14zzr/files/requirements.txt b/wandb/run-20250916_222834-l5x14zzr/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e1d19d303fe56decffe852422d8c6ee7d73d2e53 --- /dev/null +++ b/wandb/run-20250916_222834-l5x14zzr/files/requirements.txt @@ -0,0 +1,142 @@ +archer==0.1.0 +archer==0.1.0 +pip==25.2 +thinc==8.3.6 +archer==0.1.0 +torchmetrics==1.8.2 +pydantic==2.11.9 +nvidia-ml-py==13.580.82 +nvidia-nvjitlink-cu12==12.8.93 +h11==0.16.0 +anyio==4.10.0 +psutil==7.0.0 +blis==1.3.0 +nvidia-cufile-cu12==1.13.1.3 +jericho==3.3.0 +charset-normalizer==3.4.3 +wrapt==1.17.3 +Pygments==2.19.2 +lightning==2.5.5 +joblib==1.5.2 +huggingface-hub==0.34.4 +PyYAML==6.0.2 +packaging==25.0 +textblob==0.19.0 +wheel==0.45.1 +distro==1.9.0 +annotated-types==0.7.0 +nvidia-cublas-cu12==12.8.4.1 +idna==3.10 +sympy==1.14.0 +typer==0.17.4 +ninja==1.13.0 +termcolor==3.1.0 +smart_open==7.3.1 +httpcore==1.0.9 +tqdm==4.67.1 +urllib3==2.5.0 +lightning-utilities==0.15.2 +accelerate==1.10.1 +hjson==3.1.0 +networkx==3.4.2 +tokenizers==0.22.0 +rich==14.1.0 +spacy==3.8.7 +attrs==25.3.0 +cloudpathlib==0.22.0 +propcache==0.3.2 +triton==3.4.0 +setuptools==78.1.1 +certifi==2025.8.3 +gitdb==4.0.12 +nvidia-cuda-nvrtc-cu12==12.8.93 +preshed==3.0.10 +py-cpuinfo==9.0.0 +confection==0.1.5 +nvidia-cusolver-cu12==11.7.3.90 +wasabi==1.1.3 +Jinja2==3.1.6 +bitsandbytes==0.47.0 +langcodes==3.5.0 +nvidia-curand-cu12==10.3.9.90 +aiohttp==3.12.15 +nvidia-cudnn-cu12==9.10.2.21 +pytorch-lightning==2.5.5 +filelock==3.19.1 +shellingham==1.5.4 +cymem==2.0.11 +spacy-loggers==1.0.5 +sniffio==1.3.1 +regex==2025.9.1 +aiohappyeyeballs==2.6.1 +catalogue==2.0.10 +nvidia-cuda-cupti-cu12==12.8.90 +wandb==0.21.4 +srsly==2.5.1 +jiter==0.10.0 +markdown-it-py==4.0.0 +nvidia-nvtx-cu12==12.8.90 +weasel==0.4.1 +GitPython==3.1.45 +click==8.2.1 +language_data==1.3.0 +fsspec==2025.9.0 +sentry-sdk==2.37.1 +mdurl==0.1.2 +safetensors==0.6.2 +openai==1.107.2 +nvidia-nccl-cu12==2.27.3 +einops==0.8.1 +MarkupSafe==3.0.2 +murmurhash==1.0.13 +peft==0.17.1 +sentencepiece==0.2.1 +typing-inspection==0.4.1 +httpx==0.28.1 +async-timeout==5.0.1 +platformdirs==4.4.0 +spacy-legacy==3.0.12 +typeshed_client==2.8.2 +nvidia-cusparse-cu12==12.5.8.93 +transformers==4.56.1 +typing_extensions==4.15.0 +jsonargparse==4.41.0 +numpy==2.2.6 +deepspeed==0.17.5 +frozenlist==1.7.0 +aiosignal==1.4.0 +nvidia-cusparselt-cu12==0.7.1 +msgpack==1.1.1 +pydantic_core==2.33.2 +nltk==3.9.1 +smmap==5.0.2 +CD==1.1.0 +docstring_parser==0.17.0 +multidict==6.6.4 +importlib_resources==6.5.2 +torch==2.8.0 +nvidia-cuda-runtime-cu12==12.8.90 +protobuf==6.32.1 +exceptiongroup==1.3.0 +hf-xet==1.1.10 +marisa-trie==1.3.1 +yarl==1.20.1 +requests==2.32.5 +nvidia-cufft-cu12==11.3.3.83 +mpmath==1.3.0 +inflect==7.3.1 +typeguard==4.3.0 +jaraco.collections==5.1.0 +backports.tarfile==1.2.0 +autocommand==2.2.2 +jaraco.context==5.3.0 +wheel==0.45.1 +zipp==3.19.2 +tomli==2.0.1 +platformdirs==4.2.2 +jaraco.text==3.12.1 +jaraco.functools==4.0.1 +more-itertools==10.3.0 +importlib_metadata==8.0.0 +typing_extensions==4.12.2 +packaging==24.2 diff --git a/wandb/run-20250916_222834-l5x14zzr/files/wandb-metadata.json b/wandb/run-20250916_222834-l5x14zzr/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..aa10c569c8d57cddaf5691453f5387436fc3a568 --- /dev/null +++ b/wandb/run-20250916_222834-l5x14zzr/files/wandb-metadata.json @@ -0,0 +1,58 @@ +{ + "os": "Linux-6.8.0-60-generic-x86_64-with-glibc2.35", + "python": "CPython 3.10.0", + "startedAt": "2025-09-16T14:28:34.221702Z", + "args": [ + "fit", + "--data=WordTaboo", + "--data.batch_size=2", + "--data.base_model=Qwen3-14B", + "--data.n_traj_eval=4", + "--model=OfflineArcher", + "--model.optimize_critic=True", + "--model.actor_lr=1e-5", + "--model.critic_lr=1e-5", + "--model.discount_factor=0.99", + "--model.tau=0.05", + "--model.critic_expectile=0.9", + "--model.inv_temp=1.0", + "--model.accumulate_grad_batches=4", + "--model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model", + "--trainer.fast_dev_run=False", + "--trainer.max_epoch=1", + "--trainer.logger=WandbLogger", + "--trainer.logger.init_args.project=WordTaboo-Official", + "--trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word", + "--trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0", + "--trainer.strategy=deepspeed_stage_3", + "--trainer.devices=4", + "--trainer.accelerator=gpu", + "--trainer.precision=bf16", + "--trainer.enable_model_summary=false", + "--trainer.val_check_interval=null", + "--trainer.limit_val_batches=0" + ], + "program": "/home/jiashuo/codes/OfflineArcher/main.py", + "codePath": "main.py", + "codePathLocal": "main.py", + "git": { + "remote": "git@github.com:andreazanette/OfflineArcher.git", + "commit": "56d3b72665671be853bcbc961e7cd14459a3a1d1" + }, + "email": "398128267@qq.com", + "root": ".", + "host": "somea6k", + "executable": "/home/jiashuo/anaconda3/envs/archer/bin/python3.10", + "cpu_count": 64, + "cpu_count_logical": 128, + "disk": { + "/": { + "total": "3778773131264", + "used": "3576956571648" + } + }, + "memory": { + "total": "540608565248" + }, + "writerId": "6gt0ht14c42jvi0ulqzxeutysfn6ntty" +} \ No newline at end of file diff --git a/wandb/run-20250916_222834-l5x14zzr/logs/debug-core.log b/wandb/run-20250916_222834-l5x14zzr/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..64b46314839dda73381f71b61950e33f149432a2 --- /dev/null +++ b/wandb/run-20250916_222834-l5x14zzr/logs/debug-core.log @@ -0,0 +1,8 @@ +{"time":"2025-09-16T22:28:34.256989835+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp2rcjw_i2/port-112008.txt","pid":112008,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false,"enable-dcgm-profiling":false} +{"time":"2025-09-16T22:28:34.257308693+08:00","level":"WARN","msg":"server/listeners: couldn't open Unix socket in os.TempDir()","error":"server/listeners: failed to make tempdir for Unix socket: stat /home/jiashuo/tmp: no such file or directory"} +{"time":"2025-09-16T22:28:34.258264954+08:00","level":"INFO","msg":"server: will exit if parent process dies","ppid":112008} +{"time":"2025-09-16T22:28:34.258230092+08:00","level":"INFO","msg":"server: accepting connections","addr":{"Name":"/tmp/wandb-112008-113147-1463554501/socket","Net":"unix"}} +{"time":"2025-09-16T22:28:34.446390325+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"1(@)"} +{"time":"2025-09-16T22:28:34.45460064+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"l5x14zzr","id":"1(@)"} +{"time":"2025-09-16T22:28:34.938413884+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"l5x14zzr","id":"1(@)"} +{"time":"2025-09-16T22:31:20.101959055+08:00","level":"INFO","msg":"server: parent process exited, terminating service process"} diff --git a/wandb/run-20250916_222834-l5x14zzr/logs/debug-internal.log b/wandb/run-20250916_222834-l5x14zzr/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0c55927834b22170cc513aaf37a398d3153d9551 --- /dev/null +++ b/wandb/run-20250916_222834-l5x14zzr/logs/debug-internal.log @@ -0,0 +1,7 @@ +{"time":"2025-09-16T22:28:34.454782895+08:00","level":"INFO","msg":"stream: starting","core version":"0.21.4"} +{"time":"2025-09-16T22:28:34.480100793+08:00","level":"ERROR","msg":"monitor: failed to initialize GPU resource: monitor: could not create portfile"} +{"time":"2025-09-16T22:28:34.938264456+08:00","level":"INFO","msg":"stream: created new stream","id":"l5x14zzr"} +{"time":"2025-09-16T22:28:34.93839702+08:00","level":"INFO","msg":"stream: started","id":"l5x14zzr"} +{"time":"2025-09-16T22:28:34.938433109+08:00","level":"INFO","msg":"writer: started","stream_id":"l5x14zzr"} +{"time":"2025-09-16T22:28:34.93862102+08:00","level":"INFO","msg":"sender: started","stream_id":"l5x14zzr"} +{"time":"2025-09-16T22:28:34.938722342+08:00","level":"INFO","msg":"handler: started","stream_id":"l5x14zzr"} diff --git a/wandb/run-20250916_222834-l5x14zzr/logs/debug.log b/wandb/run-20250916_222834-l5x14zzr/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..af0ca294e9248146c2a3ca76a837e626f498ee01 --- /dev/null +++ b/wandb/run-20250916_222834-l5x14zzr/logs/debug.log @@ -0,0 +1,22 @@ +2025-09-16 22:28:34,225 INFO MainThread:112008 [wandb_setup.py:_flush():81] Current SDK version is 0.21.4 +2025-09-16 22:28:34,225 INFO MainThread:112008 [wandb_setup.py:_flush():81] Configure stats pid to 112008 +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/.config/wandb/settings +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_setup.py:_flush():81] Loading settings from /home/jiashuo/codes/OfflineArcher/wandb/settings +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_setup.py:_flush():81] Loading settings from environment variables +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_init.py:setup_run_log_directory():686] Logging user logs to ./wandb/run-20250916_222834-l5x14zzr/logs/debug.log +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_init.py:setup_run_log_directory():687] Logging internal logs to ./wandb/run-20250916_222834-l5x14zzr/logs/debug-internal.log +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_init.py:init():813] calling init triggers +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_init.py:init():818] wandb.init called with sweep_config: {} +config: {'_wandb': {}} +2025-09-16 22:28:34,226 INFO MainThread:112008 [wandb_init.py:init():854] starting backend +2025-09-16 22:28:34,446 INFO MainThread:112008 [wandb_init.py:init():857] sending inform_init request +2025-09-16 22:28:34,451 INFO MainThread:112008 [wandb_init.py:init():865] backend started and connected +2025-09-16 22:28:34,454 INFO MainThread:112008 [wandb_init.py:init():936] updated telemetry +2025-09-16 22:28:34,467 INFO MainThread:112008 [wandb_init.py:init():960] communicating run to backend with 90.0 second timeout +2025-09-16 22:28:35,322 INFO MainThread:112008 [wandb_init.py:init():1011] starting run threads in backend +2025-09-16 22:28:35,497 INFO MainThread:112008 [wandb_run.py:_console_start():2506] atexit reg +2025-09-16 22:28:35,498 INFO MainThread:112008 [wandb_run.py:_redirect():2354] redirect: wrap_raw +2025-09-16 22:28:35,498 INFO MainThread:112008 [wandb_run.py:_redirect():2423] Wrapping output streams. +2025-09-16 22:28:35,499 INFO MainThread:112008 [wandb_run.py:_redirect():2446] Redirects installed. +2025-09-16 22:28:35,504 INFO MainThread:112008 [wandb_init.py:init():1049] run started, returning control to user process +2025-09-16 22:29:44,749 INFO MainThread:112008 [wandb_run.py:_config_callback():1392] config_cb None None {'model_name_or_path': '/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model', 'inv_temp': 1.0, 'actor_lr': 1e-05, 'critic_lr': 1e-05, 'tau': 0.05, 'accumulate_grad_batches': 4, 'discount_factor': 0.99, 'critic_expectile': 0.9, 'optimize_critic': True, 'actor_checkpoint': None, 'critic_checkpoint': None, '_instantiator': 'lightning.pytorch.cli.instantiate_module'} diff --git a/wandb/run-20250916_222834-l5x14zzr/run-l5x14zzr.wandb b/wandb/run-20250916_222834-l5x14zzr/run-l5x14zzr.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250916_223209-ieelz6k1/files/config.yaml b/wandb/run-20250916_223209-ieelz6k1/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ce267b3d3077b9d2c4c1b2c416ec2a3b8dbbf5f9 --- /dev/null +++ b/wandb/run-20250916_223209-ieelz6k1/files/config.yaml @@ -0,0 +1,115 @@ +_instantiator: + value: lightning.pytorch.cli.instantiate_module +_wandb: + value: + cli_version: 0.21.4 + e: + bvo4u1ny3q4vifd5gbtrs2gude5sn3i4: + args: + - fit + - --data=WordTaboo + - --data.batch_size=2 + - --data.base_model=Qwen3-14B + - --data.n_traj_eval=4 + - --model=OfflineArcher + - --model.optimize_critic=True + - --model.actor_lr=1e-5 + - --model.critic_lr=1e-5 + - --model.discount_factor=0.99 + - --model.tau=0.05 + - --model.critic_expectile=0.9 + - --model.inv_temp=1.0 + - --model.accumulate_grad_batches=4 + - --model.model_name_or_path=/home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model + - --trainer.fast_dev_run=False + - --trainer.max_epoch=1 + - --trainer.logger=WandbLogger + - --trainer.logger.init_args.project=WordTaboo-Official + - --trainer.default_root_dir=checkpoints/archer_Qwen3-14B_word + - --trainer.logger.init_args.name=Test-AC-critic_expectile_0.9-inv_temp_1.0 + - --trainer.strategy=deepspeed_stage_3 + - --trainer.devices=4 + - --trainer.accelerator=gpu + - --trainer.precision=bf16 + - --trainer.enable_model_summary=false + - --trainer.val_check_interval=null + - --trainer.limit_val_batches=0 + codePath: main.py + codePathLocal: main.py + cpu_count: 64 + cpu_count_logical: 128 + disk: + /: + total: "3778773131264" + used: "3576957009920" + email: 398128267@qq.com + executable: /home/jiashuo/anaconda3/envs/archer/bin/python3.10 + git: + commit: 56d3b72665671be853bcbc961e7cd14459a3a1d1 + remote: git@github.com:andreazanette/OfflineArcher.git + host: somea6k + memory: + total: "540608565248" + os: Linux-6.8.0-60-generic-x86_64-with-glibc2.35 + program: /home/jiashuo/codes/OfflineArcher/main.py + python: CPython 3.10.0 + root: . + startedAt: "2025-09-16T14:32:09.461995Z" + writerId: bvo4u1ny3q4vifd5gbtrs2gude5sn3i4 + m: + - "1": trainer/global_step + "6": + - 3 + "7": [] + - "2": '*' + "5": 1 + "6": + - 1 + "7": [] + python_version: 3.10.0 + t: + "1": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "2": + - 1 + - 11 + - 49 + - 71 + - 98 + - 106 + "3": + - 7 + - 13 + - 66 + "4": 3.10.0 + "5": 0.21.4 + "6": 4.56.1 + "12": 0.21.4 + "13": linux-x86_64 +accumulate_grad_batches: + value: 4 +actor_checkpoint: + value: null +actor_lr: + value: 1e-05 +critic_checkpoint: + value: null +critic_expectile: + value: 0.9 +critic_lr: + value: 1e-05 +discount_factor: + value: 0.99 +inv_temp: + value: 1 +model_name_or_path: + value: /home/jiashuo/codes/ForesightOptim/checkpoints/im_Qwen3-14B_word/merged_model +optimize_critic: + value: true +tau: + value: 0.05 diff --git a/wandb/run-20250916_223209-ieelz6k1/files/output.log b/wandb/run-20250916_223209-ieelz6k1/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c89341d32bbe97834d805691e432be7a122bfafb --- /dev/null +++ b/wandb/run-20250916_223209-ieelz6k1/files/output.log @@ -0,0 +1,11 @@ +The length of the dataset is: 13220 + + *** Dataset Trimming Now Disabled. Please Called the Subroutine for triming +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/callbacks/model_checkpoint.py:751: Checkpoint directory /home/jiashuo/codes/OfflineArcher/models exists and is not empty. +Enabling DeepSpeed BF16. Model parameters and inputs will be cast to `bfloat16`. +LOCAL_RANK: 0 - CUDA_VISIBLE_DEVICES: [4,5,6,7] +Parameter Offload - Persistent parameters statistics: param_count = 601, numel = 5932040 +/home/jiashuo/anaconda3/envs/archer/lib/python3.10/site-packages/lightning/pytorch/loops/fit_loop.py:527: Found 1063 module(s) in eval mode at the start of training. This may lead to unexpected behavior during training. If this is intentional, you can ignore this warning. +Epoch 0: 0%| | 0/1652 [00:00= max_turns * 2: + break + + return "tied game", history_length + + +# djw 20250812 目前没有使用这个函数 +def convert_game_history_to_query(history, target, max_turns=5): + # Select the first game rule prompt and the first player instruct prompt + GAME_RULE_PROMPT = GAME_RULE_PROMPTS[0] + + history_str = "" + for i, message in enumerate(history): + history_str += "\n - {}: {}".format(message["role"], message["content"]) + + if len(history) == 0: + query = ( + GAME_RULE_PROMPT.format(max_turns=max_turns) + + "The game is just initialized." + ) + next_player = ROLES[0] + + else: + query = ( + GAME_RULE_PROMPT.format(max_turns=max_turns) + + "\n### Game History:" + + history_str + ) + if history[-1]["role"] == ROLES[0]: + next_player = ROLES[1] + else: + next_player = ROLES[0] + + query += INSTRUCT_PROMPTS[next_player].format(target=target) + return query + + +# djw 20250812 使用的是这个函数 +def randomly_convert_game_history_to_query(history, target, max_turns=5, **kwargs): + if "random_seed" in kwargs: + random.seed(kwargs["random_seed"]) + + if len(history) == 0: + next_player = ROLES[0] + else: + if history[-1]["role"] == ROLES[0]: + next_player = ROLES[1] + else: + next_player = ROLES[0] + + dialog_prefix = "\n" + random.choice( + ["\n - ", "\n### ", "\n## ", "\n# ", "\n *** ", "\n **", "\n\n"] + ) + answer_str, question_str = random.choice( + [ + (next_player, ROLES[1] if next_player == ROLES[0] else ROLES[0]), + ("Assistant", "Human"), + ("Answer", "Question"), + ("Response", "Query"), + ("A", "Q"), + ] + ) + + player_prefix = { + ROLES[0]: answer_str if next_player == ROLES[0] else question_str, + ROLES[1]: answer_str if next_player == ROLES[1] else question_str, + } + + history_str = "" + for i, message in enumerate(history): + history_str += "{}{}: {}".format( + dialog_prefix, player_prefix[message["role"]], message["content"] + ) + + prompt_type = random.choice(["chat", "chat_inverse", "alpaca"]) + system_prefix = random.choice(["Rules", "Game Rule", "System"]) + + GAME_RULE_PROMPT = random.choice(GAME_RULE_PROMPTS) + system_prompt = GAME_RULE_PROMPT.format(max_turns=max_turns) + + if "chat" in prompt_type: + system_prompt += "\n\n" + PLAYER_INSTRUCT_PROMPTS[next_player].format( + target=target + ) + + if len(history) == 0: + history_str = "" + system_prompt += "The game is just initialized. " + + system_str = f"{dialog_prefix}{system_prefix}: {system_prompt}" + if "inverse" in prompt_type: + query = ( + history_str + + system_str + + dialog_prefix + + player_prefix[next_player] + + ": " + ) + else: + query = ( + system_str + + history_str + + dialog_prefix + + player_prefix[next_player] + + ": " + ) + + elif prompt_type == "alpaca": + if random.uniform(0, 1) < 0.2: + system_prompt = system_prefix + ": " + system_prompt + + if len(history) == 0: + query = system_prompt + "The game is just initialized. " + else: + query = ( + system_prompt + dialog_prefix + "Game History:" + history_str + "\n\n" + ) + + if random.uniform(0, 1) < 0.2: + query += ( + PLAYER_INSTRUCT_PROMPTS[next_player].format(target=target)[:-1] + ": " + ) + else: + query += ( + PLAYER_INSTRUCT_PROMPTS[next_player].format(target=target) + + dialog_prefix + + player_prefix[next_player] + + ": " + ) + + return query