youssefedweqd commited on
Commit
f321413
·
verified ·
1 Parent(s): a3b04b2

Training in progress, step 18100, checkpoint

Browse files
last-checkpoint/adapter_config.json CHANGED
@@ -25,12 +25,12 @@
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
28
- "o_proj",
29
- "gate_proj",
30
- "down_proj",
31
  "v_proj",
 
 
32
  "k_proj",
33
- "up_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
 
25
  "revision": null,
26
  "target_modules": [
27
  "q_proj",
28
+ "up_proj",
 
 
29
  "v_proj",
30
+ "gate_proj",
31
+ "o_proj",
32
  "k_proj",
33
+ "down_proj"
34
  ],
35
  "task_type": "CAUSAL_LM",
36
  "trainable_token_indices": null,
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f91e3c5f21cf129551bcf7c69bf4d8d16a4ef7f114f3eb09634873a47bacc359
3
  size 161533160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0503d72fbbe734bd1aa1b8c5ab147618e57fdf12c863564129dd2fd420090e10
3
  size 161533160
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4014d31deb4cce624d0816b7ea019aa4bc828b0109dcbd09f7797bdcac508304
3
  size 323292202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce7e3b0e28911f6b755418ff9c22e6d5a076b7ea9b6a3985fb3ac3d87c247dd4
3
  size 323292202
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6863575074205cdae5bd9a3add89cf69a3951c4f077fb3f56f28f0fe9457c14f
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8463e4bd1b3065b876119025f33b75a288e72e59b85af3eab3c7b7c3838477e6
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -2,9 +2,9 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 2.7152317880794703,
6
  "eval_steps": 100,
7
- "global_step": 16400,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -2336,6 +2336,244 @@
2336
  "learning_rate": 2.7228170036058153e-06,
2337
  "loss": 0.2617,
2338
  "step": 16400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2339
  }
2340
  ],
2341
  "logging_steps": 50,
@@ -2355,7 +2593,7 @@
2355
  "attributes": {}
2356
  }
2357
  },
2358
- "total_flos": 1.2874605314388787e+17,
2359
  "train_batch_size": 1,
2360
  "trial_name": null,
2361
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 2.9966887417218544,
6
  "eval_steps": 100,
7
+ "global_step": 18100,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
2336
  "learning_rate": 2.7228170036058153e-06,
2337
  "loss": 0.2617,
2338
  "step": 16400
2339
+ },
2340
+ {
2341
+ "epoch": 2.7235099337748343,
2342
+ "grad_norm": 1.3112074136734009,
2343
+ "learning_rate": 2.5682530093579626e-06,
2344
+ "loss": 0.2751,
2345
+ "step": 16450
2346
+ },
2347
+ {
2348
+ "epoch": 2.7317880794701987,
2349
+ "grad_norm": 1.3641598224639893,
2350
+ "learning_rate": 2.418089536248769e-06,
2351
+ "loss": 0.2741,
2352
+ "step": 16500
2353
+ },
2354
+ {
2355
+ "epoch": 2.7400662251655628,
2356
+ "grad_norm": 1.2237251996994019,
2357
+ "learning_rate": 2.2723405158236455e-06,
2358
+ "loss": 0.2772,
2359
+ "step": 16550
2360
+ },
2361
+ {
2362
+ "epoch": 2.748344370860927,
2363
+ "grad_norm": 1.171391248703003,
2364
+ "learning_rate": 2.1310194700733697e-06,
2365
+ "loss": 0.2715,
2366
+ "step": 16600
2367
+ },
2368
+ {
2369
+ "epoch": 2.756622516556291,
2370
+ "grad_norm": 1.3004522323608398,
2371
+ "learning_rate": 1.9941395101795655e-06,
2372
+ "loss": 0.2712,
2373
+ "step": 16650
2374
+ },
2375
+ {
2376
+ "epoch": 2.7649006622516556,
2377
+ "grad_norm": 1.0566428899765015,
2378
+ "learning_rate": 1.861713335298282e-06,
2379
+ "loss": 0.2758,
2380
+ "step": 16700
2381
+ },
2382
+ {
2383
+ "epoch": 2.77317880794702,
2384
+ "grad_norm": 1.5244101285934448,
2385
+ "learning_rate": 1.7337532313818294e-06,
2386
+ "loss": 0.2786,
2387
+ "step": 16750
2388
+ },
2389
+ {
2390
+ "epoch": 2.781456953642384,
2391
+ "grad_norm": 1.2633907794952393,
2392
+ "learning_rate": 1.6102710700389656e-06,
2393
+ "loss": 0.272,
2394
+ "step": 16800
2395
+ },
2396
+ {
2397
+ "epoch": 2.789735099337748,
2398
+ "grad_norm": 1.2915066480636597,
2399
+ "learning_rate": 1.4912783074334469e-06,
2400
+ "loss": 0.2839,
2401
+ "step": 16850
2402
+ },
2403
+ {
2404
+ "epoch": 2.7980132450331126,
2405
+ "grad_norm": 1.2640457153320312,
2406
+ "learning_rate": 1.3767859832212016e-06,
2407
+ "loss": 0.2756,
2408
+ "step": 16900
2409
+ },
2410
+ {
2411
+ "epoch": 2.806291390728477,
2412
+ "grad_norm": 1.3004051446914673,
2413
+ "learning_rate": 1.2668047195261379e-06,
2414
+ "loss": 0.2747,
2415
+ "step": 16950
2416
+ },
2417
+ {
2418
+ "epoch": 2.814569536423841,
2419
+ "grad_norm": 1.3884724378585815,
2420
+ "learning_rate": 1.1613447199546158e-06,
2421
+ "loss": 0.2707,
2422
+ "step": 17000
2423
+ },
2424
+ {
2425
+ "epoch": 2.8228476821192054,
2426
+ "grad_norm": 1.416150689125061,
2427
+ "learning_rate": 1.0604157686488313e-06,
2428
+ "loss": 0.2702,
2429
+ "step": 17050
2430
+ },
2431
+ {
2432
+ "epoch": 2.8311258278145695,
2433
+ "grad_norm": 1.3139573335647583,
2434
+ "learning_rate": 9.640272293790998e-07,
2435
+ "loss": 0.2721,
2436
+ "step": 17100
2437
+ },
2438
+ {
2439
+ "epoch": 2.839403973509934,
2440
+ "grad_norm": 1.2322123050689697,
2441
+ "learning_rate": 8.721880446750708e-07,
2442
+ "loss": 0.2734,
2443
+ "step": 17150
2444
+ },
2445
+ {
2446
+ "epoch": 2.847682119205298,
2447
+ "grad_norm": 1.3550212383270264,
2448
+ "learning_rate": 7.849067349961381e-07,
2449
+ "loss": 0.2757,
2450
+ "step": 17200
2451
+ },
2452
+ {
2453
+ "epoch": 2.8559602649006623,
2454
+ "grad_norm": 1.4651058912277222,
2455
+ "learning_rate": 7.02191397940899e-07,
2456
+ "loss": 0.274,
2457
+ "step": 17250
2458
+ },
2459
+ {
2460
+ "epoch": 2.8642384105960264,
2461
+ "grad_norm": 1.3420034646987915,
2462
+ "learning_rate": 6.240497074959162e-07,
2463
+ "loss": 0.2721,
2464
+ "step": 17300
2465
+ },
2466
+ {
2467
+ "epoch": 2.872516556291391,
2468
+ "grad_norm": 1.2362934350967407,
2469
+ "learning_rate": 5.504889133237656e-07,
2470
+ "loss": 0.279,
2471
+ "step": 17350
2472
+ },
2473
+ {
2474
+ "epoch": 2.880794701986755,
2475
+ "grad_norm": 1.1310467720031738,
2476
+ "learning_rate": 4.815158400904185e-07,
2477
+ "loss": 0.2744,
2478
+ "step": 17400
2479
+ },
2480
+ {
2481
+ "epoch": 2.8890728476821192,
2482
+ "grad_norm": 1.2838894128799438,
2483
+ "learning_rate": 4.1713688683208686e-07,
2484
+ "loss": 0.2795,
2485
+ "step": 17450
2486
+ },
2487
+ {
2488
+ "epoch": 2.8973509933774837,
2489
+ "grad_norm": 1.2171522378921509,
2490
+ "learning_rate": 3.573580263615539e-07,
2491
+ "loss": 0.2778,
2492
+ "step": 17500
2493
+ },
2494
+ {
2495
+ "epoch": 2.9056291390728477,
2496
+ "grad_norm": 0.9482976198196411,
2497
+ "learning_rate": 3.0218480471403965e-07,
2498
+ "loss": 0.2757,
2499
+ "step": 17550
2500
+ },
2501
+ {
2502
+ "epoch": 2.9139072847682117,
2503
+ "grad_norm": 1.2617802619934082,
2504
+ "learning_rate": 2.5162234063264567e-07,
2505
+ "loss": 0.274,
2506
+ "step": 17600
2507
+ },
2508
+ {
2509
+ "epoch": 2.922185430463576,
2510
+ "grad_norm": 1.1305478811264038,
2511
+ "learning_rate": 2.0567532509348508e-07,
2512
+ "loss": 0.2701,
2513
+ "step": 17650
2514
+ },
2515
+ {
2516
+ "epoch": 2.9304635761589406,
2517
+ "grad_norm": 1.1746705770492554,
2518
+ "learning_rate": 1.6434802087046397e-07,
2519
+ "loss": 0.2693,
2520
+ "step": 17700
2521
+ },
2522
+ {
2523
+ "epoch": 2.9387417218543046,
2524
+ "grad_norm": 1.2440382242202759,
2525
+ "learning_rate": 1.276442621397922e-07,
2526
+ "loss": 0.2637,
2527
+ "step": 17750
2528
+ },
2529
+ {
2530
+ "epoch": 2.9470198675496686,
2531
+ "grad_norm": 0.7597376704216003,
2532
+ "learning_rate": 9.556745412425683e-08,
2533
+ "loss": 0.2657,
2534
+ "step": 17800
2535
+ },
2536
+ {
2537
+ "epoch": 2.955298013245033,
2538
+ "grad_norm": 1.2113037109375,
2539
+ "learning_rate": 6.812057277733042e-08,
2540
+ "loss": 0.272,
2541
+ "step": 17850
2542
+ },
2543
+ {
2544
+ "epoch": 2.9635761589403975,
2545
+ "grad_norm": 1.07718825340271,
2546
+ "learning_rate": 4.530616450704184e-08,
2547
+ "loss": 0.2696,
2548
+ "step": 17900
2549
+ },
2550
+ {
2551
+ "epoch": 2.9718543046357615,
2552
+ "grad_norm": 1.29874849319458,
2553
+ "learning_rate": 2.712634593974861e-08,
2554
+ "loss": 0.2712,
2555
+ "step": 17950
2556
+ },
2557
+ {
2558
+ "epoch": 2.980132450331126,
2559
+ "grad_norm": 1.1109284162521362,
2560
+ "learning_rate": 1.358280372376064e-08,
2561
+ "loss": 0.2695,
2562
+ "step": 18000
2563
+ },
2564
+ {
2565
+ "epoch": 2.98841059602649,
2566
+ "grad_norm": 1.177689790725708,
2567
+ "learning_rate": 4.6767943728598386e-09,
2568
+ "loss": 0.277,
2569
+ "step": 18050
2570
+ },
2571
+ {
2572
+ "epoch": 2.9966887417218544,
2573
+ "grad_norm": 1.1074708700180054,
2574
+ "learning_rate": 4.091441497322546e-10,
2575
+ "loss": 0.2837,
2576
+ "step": 18100
2577
  }
2578
  ],
2579
  "logging_steps": 50,
 
2593
  "attributes": {}
2594
  }
2595
  },
2596
+ "total_flos": 1.4196983428389274e+17,
2597
  "train_batch_size": 1,
2598
  "trial_name": null,
2599
  "trial_params": null
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e7fd7045b4a15373c4d905f915edbb1436664635b44fe1df1bfa95de5188043
3
  size 5752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e16cd32295896d2f57f1f19dbc49310a6ac6c29a2ec738cf4fb0c11d34db588b
3
  size 5752