{ "model_name_or_path": "huggyllama/llama-7b", "max_length": 512, "trust_remote_code": true, "train_datasets": [ [ "alpaca", { "proportion": 1.0 } ] ], "eval_datasets": null, "epochs": 3, "per_device_train_batch_size": 4, "per_device_eval_batch_size": 4, "gradient_accumulation_steps": 16, "gradient_checkpointing": true, "lr": 2e-05, "lr_scheduler_type": "cosine", "lr_warmup_ratio": 0.03, "weight_decay": 0.0, "seed": 42, "fp16": false, "bf16": true, "tf32": true, "eval_strategy": "epoch", "eval_interval": 1000000, "need_eval": false, "eval_split_ratio": null, "output_dir": "/data/jiongxiao_wang/rlhf_attack/safe-rlhf/output/sft", "log_type": "wandb", "log_dir": "/data/jiongxiao_wang/rlhf_attack/safe-rlhf/output/sft", "log_project": "Safe-RLHF-SFT", "log_run_name": "sft-2023-12-31-20-07-40", "save_16bit": false, "save_interval": 1000000, "local_rank": 0, "zero_stage": 3, "deepspeed": false, "deepspeed_config": null, "deepscale": false, "deepscale_config": null, "deepspeed_mpi": false, "global_rank": 0, "device": { "type": "torch.device", "repr": "device(type='cuda', index=0)" }, "num_update_steps_per_epoch": 204, "total_training_steps": 612 }