mohammadhossein commited on
Commit
5694fa0
·
verified ·
1 Parent(s): ec55243

Training in progress, step 96

Browse files
adapter_config.json CHANGED
@@ -6,6 +6,8 @@
6
  },
7
  "base_model_name_or_path": "facebook/musicgen-melody",
8
  "bias": "none",
 
 
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
11
  "init_lora_weights": true,
@@ -14,6 +16,7 @@
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
  "lora_alpha": 16,
 
17
  "lora_dropout": 0.05,
18
  "megatron_config": null,
19
  "megatron_core": "megatron.core",
@@ -23,21 +26,21 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
26
  "embed_tokens.1",
27
- "lm_heads.1",
28
- "q_proj",
29
  "audio_enc_to_dec_proj",
30
- "embed_tokens.2",
 
 
 
 
31
  "embed_tokens.3",
 
32
  "lm_heads.2",
33
- "fc2",
34
- "out_proj",
35
- "lm_heads.3",
36
  "lm_heads.0",
37
  "k_proj",
38
- "fc1",
39
- "embed_tokens.0",
40
- "v_proj",
41
  "enc_to_dec_proj"
42
  ],
43
  "task_type": null,
 
6
  },
7
  "base_model_name_or_path": "facebook/musicgen-melody",
8
  "bias": "none",
9
+ "eva_config": null,
10
+ "exclude_modules": null,
11
  "fan_in_fan_out": false,
12
  "inference_mode": true,
13
  "init_lora_weights": true,
 
16
  "layers_to_transform": null,
17
  "loftq_config": {},
18
  "lora_alpha": 16,
19
+ "lora_bias": false,
20
  "lora_dropout": 0.05,
21
  "megatron_config": null,
22
  "megatron_core": "megatron.core",
 
26
  "rank_pattern": {},
27
  "revision": null,
28
  "target_modules": [
29
+ "embed_tokens.2",
30
  "embed_tokens.1",
31
+ "lm_heads.3",
 
32
  "audio_enc_to_dec_proj",
33
+ "out_proj",
34
+ "fc2",
35
+ "embed_tokens.0",
36
+ "v_proj",
37
+ "lm_heads.1",
38
  "embed_tokens.3",
39
+ "q_proj",
40
  "lm_heads.2",
41
+ "fc1",
 
 
42
  "lm_heads.0",
43
  "k_proj",
 
 
 
44
  "enc_to_dec_proj"
45
  ],
46
  "task_type": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cac26d0ccfdc63dcf798ec16bae40b3cc424f5732f01254e0759ee23991c5ecb
3
  size 87103456
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01be36eba6a1132ecfe1d2cea4c8ecaf17a806c48b6ba678b48bed1ebd2de52c
3
  size 87103456
tokenizer.json CHANGED
@@ -1,7 +1,14 @@
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
  "truncation": null,
4
+ "padding": {
5
+ "strategy": "BatchLongest",
6
+ "direction": "Right",
7
+ "pad_to_multiple_of": null,
8
+ "pad_id": 0,
9
+ "pad_type_id": 0,
10
+ "pad_token": "<pad>"
11
+ },
12
  "added_tokens": [
13
  {
14
  "id": 0,
tokenizer_config.json CHANGED
@@ -930,6 +930,7 @@
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "max_length": null,
934
  "model_max_length": 512,
935
  "pad_to_multiple_of": null,
 
930
  "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
933
+ "extra_special_tokens": {},
934
  "max_length": null,
935
  "model_max_length": 512,
936
  "pad_to_multiple_of": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c9696e5beb9bf03a4819ceb7582dd188a3c9b89376ea75515f2229e4b4edc9b
3
- size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:336c1331ee08cb263ad1084726e51f70f884bb6e12ddc339c37cb60fc7a19e12
3
+ size 5496