Zelyanoth commited on
Commit
c0ad23f
·
verified ·
1 Parent(s): d22f101

Training in progress, step 1800, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:532b3c8dd0fc48025330f131147e3bf655f8eb0ccc1b6f1be715ffcd37eeb0e6
3
  size 16406680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:327212e84ba8a347d12bec59ece90867c94c9182c0b2908575049152721ce4a1
3
  size 16406680
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0a018a28b363e2fe865407a2add098a16543edec7054f4d6b3acab628cd2db0b
3
  size 6573615
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17bc83c01c7d996d4bae04714be783adfcc9079cc0fc45cb1321e249c2e52693
3
  size 6573615
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d482856761d4619002198ae9b8fc4dbddbf83ef7b336054705a7e2ace9dc37f5
3
  size 14645
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:055ad8c7bb3354d2c4d9cad3fc3149048847f95be5ca999100b4b2b158d986d5
3
  size 14645
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9a24a7d5288f0f1fae49cca8a8fa22aa94481eb5acf767408b35e0d501a89b8
3
  size 1383
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e7e8108fcdb3a0486aefb31155f0b6d11fb09d6229467e1e333e5c291d89d9a
3
  size 1383
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e95c77d1c57a34d51faaae9500eeda620e307df8870032eb6ea3441b2e2dfba
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:104dbcce8836684729bcaeff6b52ebfb14715e797d21f0ca4a447252176b4567
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 1600,
3
- "best_metric": 46.76727833990739,
4
- "best_model_checkpoint": "./whisper-large-v3-turbo-swahili/checkpoint-1600",
5
- "epoch": 2.8776978417266186,
6
  "eval_steps": 200,
7
- "global_step": 1600,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -304,6 +304,43 @@
304
  "eval_steps_per_second": 0.176,
305
  "eval_wer": 46.76727833990739,
306
  "step": 1600
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
307
  }
308
  ],
309
  "logging_steps": 50,
@@ -323,7 +360,7 @@
323
  "attributes": {}
324
  }
325
  },
326
- "total_flos": 9.869390776369152e+19,
327
  "train_batch_size": 36,
328
  "trial_name": null,
329
  "trial_params": null
 
1
  {
2
+ "best_global_step": 1800,
3
+ "best_metric": 46.73297890584806,
4
+ "best_model_checkpoint": "./whisper-large-v3-turbo-swahili/checkpoint-1800",
5
+ "epoch": 3.237410071942446,
6
  "eval_steps": 200,
7
+ "global_step": 1800,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
304
  "eval_steps_per_second": 0.176,
305
  "eval_wer": 46.76727833990739,
306
  "step": 1600
307
+ },
308
+ {
309
+ "epoch": 2.9676258992805753,
310
+ "grad_norm": 0.8634299635887146,
311
+ "learning_rate": 1.3311007957559682e-05,
312
+ "loss": 0.5152,
313
+ "step": 1650
314
+ },
315
+ {
316
+ "epoch": 3.0575539568345325,
317
+ "grad_norm": 0.8060579895973206,
318
+ "learning_rate": 1.3211538461538462e-05,
319
+ "loss": 0.4906,
320
+ "step": 1700
321
+ },
322
+ {
323
+ "epoch": 3.147482014388489,
324
+ "grad_norm": 0.9339563846588135,
325
+ "learning_rate": 1.3112068965517243e-05,
326
+ "loss": 0.492,
327
+ "step": 1750
328
+ },
329
+ {
330
+ "epoch": 3.237410071942446,
331
+ "grad_norm": 0.8844484090805054,
332
+ "learning_rate": 1.3012599469496021e-05,
333
+ "loss": 0.514,
334
+ "step": 1800
335
+ },
336
+ {
337
+ "epoch": 3.237410071942446,
338
+ "eval_loss": 0.7331941723823547,
339
+ "eval_runtime": 393.5575,
340
+ "eval_samples_per_second": 1.779,
341
+ "eval_steps_per_second": 0.178,
342
+ "eval_wer": 46.73297890584806,
343
+ "step": 1800
344
  }
345
  ],
346
  "logging_steps": 50,
 
360
  "attributes": {}
361
  }
362
  },
363
+ "total_flos": 1.1101007357411328e+20,
364
  "train_batch_size": 36,
365
  "trial_name": null,
366
  "trial_params": null