| _target_: null |
| job: |
| name: testing__pvv2_lora |
| mode: local |
| dry_run: false |
| work_dir: null |
| slurm: |
| time_limit: null |
| constraint: null |
| memory: null |
| cpus_per_task: 16 |
| partition: null |
| mail_user: null |
| execution: |
| nodes: 1 |
| gpus_per_node: 2 |
| num_gpus: null |
| hostfile: null |
| secrets_file: ./secrets.env |
| model: |
| name_or_path: Qwen/Qwen2.5-1.5B-Instruct |
| finetuning_type: lora |
| dataset: |
| name: TAUR_dev_D_SFT_C_ours_cd3arg_10responses_reflections10_formats_C_full |
| dir: null |
| info_json: null |
| template: qwen |
| cutoff_len: 16192 |
| val_size: 0.0 |
| tokenized_path: /scratch/zrs2020/.cache/llamafactory/tokenized/TAUR_dev_D_SFT_C_ours_cd3arg_10responses_reflections10_formats_C_full |
| hf_hub_url: TAUR-dev/D-SFT_C-ours_cd3arg_10responses_reflections10_formats-C_full |
| formatting: sharegpt |
| ranking: false |
| subset: null |
| split: train |
| folder: null |
| num_samples: null |
| columns: |
| prompt: null |
| query: null |
| response: null |
| history: null |
| messages: conversations |
| system: null |
| tools: null |
| images: null |
| videos: null |
| audios: null |
| chosen: null |
| rejected: null |
| kto_tag: null |
| tags: |
| role: role |
| content: content |
| user: user |
| assistant: assistant |
| observation: null |
| function: null |
| system: null |
| output: |
| experiment_dir: ./experiments |
| wandb: |
| project: null |
| run_name: testing__pvv2_lora |
| entity: null |
| hf: |
| repo_id: TAUR-dev/testing__pvv2_lora |
| private: false |
| upload_artifacts: true |
| cleanup: |
| checkpoints: false |
| merged: false |
| training: |
| stage: sft |
| do_train: true |
| max_samples: 100000 |
| do_eval: false |
| save_strategy: steps |
| save_steps: 5 |
| logging_steps: 10 |
| fp16: false |
| bf16: true |
| adam_beta1: 0.9 |
| adam_beta2: 0.95 |
| overwrite_output_dir: true |
| per_device_train_batch_size: 1 |
| gradient_accumulation_steps: 1 |
| gradient_checkpointing: true |
| learning_rate: 1.0e-06 |
| lr_scheduler_type: cosine |
| num_train_epochs: 2 |
| warmup_ratio: 0.05 |
| weight_decay: 0.0001 |
| template: qwen |
| max_steps: 10 |
| preprocessing_num_workers: 16 |
| overwrite_cache: true |
| finetuning: |
| training: |
| stage: sft |
| do_train: true |
| finetuning_type: lora |
| lora_rank: 8 |
| lora_alpha: 16 |
| lora_dropout: 0.05 |
| lora_target: all |
| overwrite_cache: true |
| preprocessing_num_workers: 16 |
| dataloader_num_workers: 4 |
| logging_steps: 10 |
| save_steps: 500 |
| plot_loss: true |
| overwrite_output_dir: true |
| save_only_model: false |
| report_to: none |
| per_device_train_batch_size: 1 |
| gradient_accumulation_steps: 8 |
| learning_rate: 0.0001 |
| num_train_epochs: 3.0 |
| lr_scheduler_type: cosine |
| warmup_ratio: 0.1 |
| bf16: true |
| ddp_timeout: 180000000 |
| resume_from_checkpoint: null |
| val_size: 0.1 |
| per_device_eval_batch_size: 1 |
| eval_strategy: steps |
| eval_steps: 500 |
| do_eval: true |
| merge: |
| export_dir: null |
| export_size: 5 |
| export_device: cpu |
| export_legacy_format: false |
|
|