bnb_cfgs: | |
bnb_4bit_compute_dtype: float16 | |
bnb_4bit_quant_type: nf4 | |
bnb_4bit_use_double_quant: true | |
load_in_4bit: true | |
load_in_8bit: false | |
use_bnb: false | |
data_cfgs: | |
eval_data_files: {} | |
eval_datasets: {} | |
eval_optional_args: [] | |
eval_size: {} | |
eval_split: {} | |
eval_subset: {} | |
eval_template: {} | |
train_data_files: {} | |
train_datasets: /aifs4su/yaodong/wenqi/projects/first-time_safety/data_annotation/data_output/safe-o1_0403/baseline_dataset | |
train_name: {} | |
train_optional_args: [] | |
train_size: {} | |
train_split: train | |
train_template: Safe_thinking | |
logger_cfgs: | |
cache_dir: {} | |
log_project: safe-o1 | |
log_run_name: sft | |
log_type: wandb | |
output_dir: /aifs4su/yaodong/wenqi/projects/first-time_safety/output_models/Qwen2.5-7B-Instruct_safe_thinking | |
save_interval: 100000 | |
lora_cfgs: | |
inference_mode: false | |
lora_alpha: 16 | |
lora_dropout: 0.1 | |
r: 16 | |
save_full_model: true | |
target_modules: | |
- q_proj | |
- v_proj | |
task_type: TaskType.CAUSAL_LM | |
use_lora: false | |
model_cfgs: | |
model_max_length: 16384 | |
model_name_or_path: /aifs4su/yaodong/wenqi/models/Qwen2.5-7B-Instruct | |
trust_remote_code: true | |
special_tokens: {} | |
train_cfgs: | |
adam_betas: | |
- 0.9 | |
- 0.95 | |
adam_epsilon: 1.0e-08 | |
bf16: true | |
ds_cfgs: ds_z3_config.json | |
epochs: 3 | |
eval_interval: 10 | |
eval_strategy: steps | |
fp16: false | |
gradient_accumulation_steps: 2 | |
gradient_checkpointing: true | |
learning_rate: 2.0e-05 | |
lr_scheduler_type: constant | |
lr_warmup_ratio: 0.03 | |
max_grad_norm: 1.0 | |
per_device_eval_batch_size: 4 | |
per_device_train_batch_size: 4 | |
seed: 42 | |
weight_decay: 0.0 | |