joaormedeiros commited on
Commit
9d847e1
·
verified ·
1 Parent(s): bb7aea5

Upload finetune_config.yml with huggingface_hub

Browse files
Files changed (1) hide show
  1. finetune_config.yml +63 -0
finetune_config.yml ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ data:
2
+ dataset_name: manual_plus_executed_prompt_variance
3
+ from_langfuse: true
4
+ method: sft
5
+ num_proc: 2
6
+ seed: 3407
7
+ split: train
8
+ test_size: 0.2
9
+ logging:
10
+ file: finetune.log
11
+ model:
12
+ adapter: null
13
+ dtype: null
14
+ load_in_4bit: true
15
+ lora:
16
+ alpha: 16
17
+ bias: none
18
+ dropout: 0
19
+ gradient_checkpointing: unsloth
20
+ loftq_config: null
21
+ r: 16
22
+ random_state: 3407
23
+ target_modules:
24
+ - q_proj
25
+ - k_proj
26
+ - v_proj
27
+ - o_proj
28
+ - gate_proj
29
+ - up_proj
30
+ - down_proj
31
+ use_rslora: false
32
+ max_seq_length: 2048
33
+ name: Qwen/Qwen2.5-7B-Instruct
34
+ save:
35
+ gguf_quantization:
36
+ - q4_k_m
37
+ - q8_0
38
+ - q5_k_m
39
+ hub_repo: joaormedeiros/qwen-2.5-lite-recursive-v5-rules-iter-1
40
+ local_dir: lora_model
41
+ private: true
42
+ push_to_hub: true
43
+ save_mode: all_training_files
44
+ test_inference:
45
+ enabled: true
46
+ input: 1, 1, 2, 3, 5, 8
47
+ instruction: Continue the fibonnaci sequence.
48
+ training:
49
+ batch_size:
50
+ eval: 2
51
+ train: 4
52
+ eval_steps: 100
53
+ gradient_accumulation_steps: 4
54
+ learning_rate: 0.0002
55
+ logging_steps: 1
56
+ lr_scheduler: linear
57
+ num_epochs: 5
58
+ optimizer: adamw_8bit
59
+ output_dir: outputs
60
+ report_to: tensorboard
61
+ seed: 3407
62
+ warmup_steps: 1
63
+ weight_decay: 0.01