dayone3nder's picture
Upload folder using huggingface_hub
8e26f61 verified
bnb_cfgs:
bnb_4bit_compute_dtype: float16
bnb_4bit_quant_type: nf4
bnb_4bit_use_double_quant: true
load_in_4bit: true
load_in_8bit: false
use_bnb: false
data_cfgs:
eval_data_files: {}
eval_datasets: {}
eval_optional_args: []
eval_size: {}
eval_split: {}
eval_subset: {}
eval_template: {}
train_data_files: {}
train_datasets: /aifs4su/yaodong/wenqi/projects/first-time_safety/data_annotation/data_output/safe-o1_0403/baseline_dataset
train_name: {}
train_optional_args: []
train_size: {}
train_split: train
train_template: Safe_thinking
logger_cfgs:
cache_dir: {}
log_project: safe-o1
log_run_name: sft
log_type: wandb
output_dir: /aifs4su/yaodong/wenqi/projects/first-time_safety/output_models/Qwen2.5-7B-Instruct_safe_thinking
save_interval: 100000
lora_cfgs:
inference_mode: false
lora_alpha: 16
lora_dropout: 0.1
r: 16
save_full_model: true
target_modules:
- q_proj
- v_proj
task_type: TaskType.CAUSAL_LM
use_lora: false
model_cfgs:
model_max_length: 16384
model_name_or_path: /aifs4su/yaodong/wenqi/models/Qwen2.5-7B-Instruct
trust_remote_code: true
special_tokens: {}
train_cfgs:
adam_betas:
- 0.9
- 0.95
adam_epsilon: 1.0e-08
bf16: true
ds_cfgs: ds_z3_config.json
epochs: 3
eval_interval: 10
eval_strategy: steps
fp16: false
gradient_accumulation_steps: 2
gradient_checkpointing: true
learning_rate: 2.0e-05
lr_scheduler_type: constant
lr_warmup_ratio: 0.03
max_grad_norm: 1.0
per_device_eval_batch_size: 4
per_device_train_batch_size: 4
seed: 42
weight_decay: 0.0