ChayanM commited on
Commit
74e28d9
·
verified ·
1 Parent(s): 5be525c

Model save

Browse files
Files changed (2) hide show
  1. README.md +58 -58
  2. model.safetensors +1 -1
README.md CHANGED
@@ -15,12 +15,12 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.0975
19
- - Rouge1: 24.6871
20
- - Rouge2: 9.5762
21
- - Rougel: 20.8694
22
- - Rougelsum: 23.5961
23
- - Gen Len: 18.9192
24
 
25
  ## Model description
26
 
@@ -40,8 +40,8 @@ More information needed
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 5e-05
43
- - train_batch_size: 12
44
- - eval_batch_size: 12
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
@@ -51,56 +51,56 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
53
  |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
54
- | 0.2412 | 1.0 | 558 | 0.1433 | 25.8229 | 12.3048 | 23.7573 | 25.4769 | 19.0 |
55
- | 0.1504 | 2.0 | 1116 | 0.1213 | 25.6896 | 12.1913 | 23.6153 | 25.3363 | 19.0 |
56
- | 0.132 | 3.0 | 1674 | 0.1099 | 26.1044 | 12.1738 | 23.7995 | 25.4389 | 19.0 |
57
- | 0.1165 | 4.0 | 2232 | 0.1018 | 24.4958 | 11.6079 | 21.9255 | 24.1601 | 19.0 |
58
- | 0.105 | 5.0 | 2790 | 0.0942 | 26.0341 | 12.2669 | 23.7171 | 25.6348 | 19.0 |
59
- | 0.0942 | 6.0 | 3348 | 0.0874 | 25.3352 | 10.6409 | 22.3911 | 24.6749 | 19.0 |
60
- | 0.0847 | 7.0 | 3906 | 0.0825 | 24.3455 | 11.1631 | 22.0193 | 23.8951 | 19.0 |
61
- | 0.0764 | 8.0 | 4464 | 0.0782 | 25.4345 | 11.7359 | 23.0169 | 24.8774 | 19.0 |
62
- | 0.0615 | 9.0 | 5022 | 0.0742 | 26.5655 | 12.5315 | 24.0201 | 26.0177 | 19.0 |
63
- | 0.0546 | 10.0 | 5580 | 0.0714 | 26.984 | 10.977 | 23.3161 | 25.9544 | 19.0 |
64
- | 0.0483 | 11.0 | 6138 | 0.0689 | 26.2815 | 11.5641 | 23.1829 | 25.4578 | 19.0 |
65
- | 0.044 | 12.0 | 6696 | 0.0663 | 25.2328 | 11.3217 | 22.4545 | 24.653 | 19.0 |
66
- | 0.0383 | 13.0 | 7254 | 0.0648 | 25.9672 | 10.9082 | 22.8064 | 25.1251 | 19.0 |
67
- | 0.0351 | 14.0 | 7812 | 0.0660 | 26.0833 | 11.0382 | 22.6573 | 25.2428 | 19.0 |
68
- | 0.0313 | 15.0 | 8370 | 0.0658 | 26.7009 | 10.9455 | 22.845 | 25.6707 | 19.0 |
69
- | 0.0276 | 16.0 | 8928 | 0.0659 | 26.1769 | 10.6049 | 22.4903 | 25.321 | 19.0 |
70
- | 0.0246 | 17.0 | 9486 | 0.0661 | 26.1478 | 10.3981 | 22.4809 | 25.1318 | 19.0 |
71
- | 0.0197 | 18.0 | 10044 | 0.0682 | 25.3438 | 10.4852 | 22.1182 | 24.5048 | 19.0 |
72
- | 0.0178 | 19.0 | 10602 | 0.0689 | 25.2217 | 9.6912 | 21.4433 | 24.1131 | 19.0 |
73
- | 0.0159 | 20.0 | 11160 | 0.0707 | 24.6521 | 9.4214 | 21.0511 | 23.6559 | 18.9771 |
74
- | 0.0148 | 21.0 | 11718 | 0.0725 | 24.6477 | 9.8393 | 21.2375 | 23.8722 | 18.9933 |
75
- | 0.0133 | 22.0 | 12276 | 0.0723 | 25.2483 | 9.9924 | 21.6189 | 24.2124 | 18.9933 |
76
- | 0.0121 | 23.0 | 12834 | 0.0741 | 24.3834 | 9.8801 | 21.0599 | 23.5916 | 18.9664 |
77
- | 0.0114 | 24.0 | 13392 | 0.0757 | 25.0727 | 9.7857 | 21.3046 | 24.0167 | 18.9758 |
78
- | 0.0103 | 25.0 | 13950 | 0.0774 | 25.1959 | 10.1108 | 21.5608 | 24.1292 | 18.9098 |
79
- | 0.0089 | 26.0 | 14508 | 0.0783 | 25.5931 | 9.9812 | 21.6953 | 24.3561 | 18.9219 |
80
- | 0.0083 | 27.0 | 15066 | 0.0793 | 24.8603 | 10.0231 | 21.2615 | 23.9145 | 18.9879 |
81
- | 0.0076 | 28.0 | 15624 | 0.0802 | 24.741 | 9.6977 | 21.112 | 23.8097 | 18.9367 |
82
- | 0.0074 | 29.0 | 16182 | 0.0812 | 24.0656 | 9.4335 | 20.6021 | 23.0172 | 18.8748 |
83
- | 0.0067 | 30.0 | 16740 | 0.0838 | 24.9923 | 9.9583 | 21.2749 | 23.9427 | 18.9556 |
84
- | 0.0063 | 31.0 | 17298 | 0.0844 | 24.8869 | 9.6309 | 21.0218 | 23.7523 | 18.8789 |
85
- | 0.0058 | 32.0 | 17856 | 0.0870 | 24.8009 | 9.9887 | 21.0596 | 23.71 | 18.9139 |
86
- | 0.0054 | 33.0 | 18414 | 0.0879 | 24.9076 | 9.663 | 21.0755 | 23.8641 | 18.9287 |
87
- | 0.0052 | 34.0 | 18972 | 0.0902 | 25.0668 | 9.5739 | 21.2282 | 23.9928 | 18.9044 |
88
- | 0.0044 | 35.0 | 19530 | 0.0908 | 25.0616 | 10.0034 | 21.6482 | 23.9978 | 18.9098 |
89
- | 0.0041 | 36.0 | 20088 | 0.0912 | 25.0681 | 10.099 | 21.4527 | 24.0219 | 18.8573 |
90
- | 0.0039 | 37.0 | 20646 | 0.0916 | 24.6263 | 9.7547 | 20.8695 | 23.5722 | 18.9367 |
91
- | 0.0037 | 38.0 | 21204 | 0.0922 | 24.6973 | 9.6421 | 21.1171 | 23.733 | 18.9435 |
92
- | 0.0034 | 39.0 | 21762 | 0.0929 | 25.3821 | 9.8435 | 21.4803 | 24.3296 | 18.8439 |
93
- | 0.0032 | 40.0 | 22320 | 0.0944 | 25.2386 | 9.9245 | 21.4207 | 24.1773 | 18.9287 |
94
- | 0.003 | 41.0 | 22878 | 0.0947 | 25.2413 | 10.0581 | 21.5136 | 24.146 | 18.9623 |
95
- | 0.0028 | 42.0 | 23436 | 0.0958 | 25.1041 | 9.8452 | 21.2494 | 24.0197 | 18.9166 |
96
- | 0.0027 | 43.0 | 23994 | 0.0960 | 24.8 | 9.932 | 21.1541 | 23.7546 | 18.9529 |
97
- | 0.0024 | 44.0 | 24552 | 0.0965 | 25.1426 | 10.0351 | 21.3824 | 24.0439 | 18.9341 |
98
- | 0.0023 | 45.0 | 25110 | 0.0964 | 24.837 | 9.7853 | 21.0592 | 23.7822 | 18.9421 |
99
- | 0.0022 | 46.0 | 25668 | 0.0968 | 24.8325 | 9.8007 | 20.976 | 23.7022 | 18.9596 |
100
- | 0.0021 | 47.0 | 26226 | 0.0972 | 24.7028 | 9.6921 | 21.0038 | 23.6658 | 18.9394 |
101
- | 0.0021 | 48.0 | 26784 | 0.0974 | 24.7233 | 9.77 | 21.0449 | 23.6333 | 18.9260 |
102
- | 0.002 | 49.0 | 27342 | 0.0977 | 24.7481 | 9.6823 | 20.9024 | 23.6522 | 18.9300 |
103
- | 0.0019 | 50.0 | 27900 | 0.0975 | 24.6871 | 9.5762 | 20.8694 | 23.5961 | 18.9192 |
104
 
105
 
106
  ### Framework versions
 
15
 
16
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.0923
19
+ - Rouge1: 25.0369
20
+ - Rouge2: 10.1572
21
+ - Rougel: 21.5244
22
+ - Rougelsum: 24.0775
23
+ - Gen Len: 18.9946
24
 
25
  ## Model description
26
 
 
40
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 5e-05
43
+ - train_batch_size: 8
44
+ - eval_batch_size: 8
45
  - seed: 42
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
53
  |:-------------:|:-----:|:-----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
54
+ | 0.253 | 1.0 | 836 | 0.1372 | 29.3958 | 12.2981 | 25.5129 | 27.9289 | 19.0 |
55
+ | 0.1361 | 2.0 | 1672 | 0.1151 | 25.8361 | 12.2894 | 23.7346 | 25.47 | 19.0 |
56
+ | 0.115 | 3.0 | 2508 | 0.1037 | 25.1859 | 11.9032 | 23.1038 | 24.8338 | 19.0 |
57
+ | 0.1027 | 4.0 | 3344 | 0.0942 | 26.0345 | 12.0324 | 23.4843 | 25.5426 | 19.0 |
58
+ | 0.0873 | 5.0 | 4180 | 0.0864 | 26.1657 | 11.685 | 23.6563 | 25.6247 | 19.0 |
59
+ | 0.0742 | 6.0 | 5016 | 0.0794 | 24.3621 | 10.5113 | 21.7192 | 23.8253 | 19.0 |
60
+ | 0.0646 | 7.0 | 5852 | 0.0740 | 24.711 | 11.194 | 22.2089 | 24.1793 | 19.0 |
61
+ | 0.0542 | 8.0 | 6688 | 0.0690 | 25.0339 | 10.8651 | 22.171 | 24.4106 | 19.0 |
62
+ | 0.046 | 9.0 | 7524 | 0.0650 | 25.0982 | 11.8399 | 22.701 | 24.623 | 18.9987 |
63
+ | 0.0386 | 10.0 | 8360 | 0.0623 | 26.2563 | 10.4715 | 22.5319 | 25.1412 | 18.9987 |
64
+ | 0.0317 | 11.0 | 9196 | 0.0591 | 26.4001 | 11.8031 | 23.1653 | 25.2856 | 18.9919 |
65
+ | 0.0273 | 12.0 | 10032 | 0.0587 | 25.6521 | 11.0174 | 22.7327 | 24.9068 | 18.9879 |
66
+ | 0.0231 | 13.0 | 10868 | 0.0583 | 26.7035 | 11.2021 | 23.0121 | 25.6384 | 18.9946 |
67
+ | 0.0195 | 14.0 | 11704 | 0.0592 | 25.5747 | 10.7424 | 22.3673 | 24.6944 | 19.0 |
68
+ | 0.0167 | 15.0 | 12540 | 0.0608 | 25.3022 | 10.163 | 21.9556 | 24.3587 | 18.9596 |
69
+ | 0.0142 | 16.0 | 13376 | 0.0614 | 25.0496 | 10.0656 | 21.7629 | 24.1094 | 18.9206 |
70
+ | 0.0119 | 17.0 | 14212 | 0.0618 | 26.0112 | 10.2519 | 22.1926 | 24.8873 | 18.8735 |
71
+ | 0.0102 | 18.0 | 15048 | 0.0653 | 25.6183 | 10.04 | 22.1136 | 24.5255 | 18.9125 |
72
+ | 0.0086 | 19.0 | 15884 | 0.0671 | 24.7352 | 9.6328 | 21.0675 | 23.7704 | 18.8694 |
73
+ | 0.0076 | 20.0 | 16720 | 0.0693 | 24.9512 | 9.6635 | 21.4761 | 23.9132 | 18.9112 |
74
+ | 0.0067 | 21.0 | 17556 | 0.0708 | 24.1732 | 9.158 | 20.3408 | 23.029 | 18.8358 |
75
+ | 0.0058 | 22.0 | 18392 | 0.0732 | 24.4503 | 9.4394 | 20.8584 | 23.4242 | 18.8035 |
76
+ | 0.0048 | 23.0 | 19228 | 0.0738 | 24.8844 | 9.9125 | 21.3509 | 23.9336 | 18.8089 |
77
+ | 0.0043 | 24.0 | 20064 | 0.0777 | 25.5401 | 10.1857 | 21.8328 | 24.4294 | 18.9058 |
78
+ | 0.0038 | 25.0 | 20900 | 0.0781 | 24.2235 | 9.0445 | 20.4463 | 23.0001 | 18.9166 |
79
+ | 0.0033 | 26.0 | 21736 | 0.0801 | 25.0127 | 9.8025 | 21.3116 | 23.9683 | 18.7308 |
80
+ | 0.0029 | 27.0 | 22572 | 0.0807 | 24.5765 | 9.6283 | 20.9556 | 23.4559 | 18.9166 |
81
+ | 0.0027 | 28.0 | 23408 | 0.0830 | 24.8389 | 9.8899 | 21.4027 | 23.9416 | 18.9233 |
82
+ | 0.0024 | 29.0 | 24244 | 0.0833 | 25.3695 | 10.162 | 21.7865 | 24.3737 | 18.7106 |
83
+ | 0.0022 | 30.0 | 25080 | 0.0832 | 24.8804 | 10.0825 | 21.4621 | 24.0326 | 18.9287 |
84
+ | 0.0021 | 31.0 | 25916 | 0.0853 | 25.0049 | 9.7036 | 21.3664 | 23.9173 | 18.9044 |
85
+ | 0.0019 | 32.0 | 26752 | 0.0855 | 25.0529 | 9.4994 | 21.2781 | 24.0076 | 18.9125 |
86
+ | 0.002 | 33.0 | 27588 | 0.0852 | 24.8417 | 9.9376 | 21.2526 | 23.8552 | 18.9031 |
87
+ | 0.0015 | 34.0 | 28424 | 0.0857 | 24.6359 | 9.5179 | 20.8941 | 23.4553 | 18.8937 |
88
+ | 0.0014 | 35.0 | 29260 | 0.0858 | 25.1156 | 10.1869 | 21.5805 | 23.9664 | 18.8156 |
89
+ | 0.0013 | 36.0 | 30096 | 0.0871 | 24.739 | 9.5548 | 21.15 | 23.749 | 18.9219 |
90
+ | 0.0011 | 37.0 | 30932 | 0.0884 | 24.774 | 9.7848 | 21.2467 | 23.833 | 18.9556 |
91
+ | 0.0011 | 38.0 | 31768 | 0.0889 | 25.2656 | 9.9796 | 21.517 | 24.1836 | 18.9462 |
92
+ | 0.0011 | 39.0 | 32604 | 0.0895 | 24.6627 | 9.3783 | 20.9288 | 23.5835 | 18.9704 |
93
+ | 0.001 | 40.0 | 33440 | 0.0906 | 25.1326 | 9.814 | 21.3593 | 24.0816 | 18.9260 |
94
+ | 0.0009 | 41.0 | 34276 | 0.0900 | 25.6889 | 10.3712 | 22.0588 | 24.695 | 18.9731 |
95
+ | 0.0008 | 42.0 | 35112 | 0.0911 | 24.6819 | 9.8307 | 21.1335 | 23.7053 | 18.9071 |
96
+ | 0.0008 | 43.0 | 35948 | 0.0905 | 24.4835 | 9.7292 | 21.017 | 23.5027 | 18.9623 |
97
+ | 0.0007 | 44.0 | 36784 | 0.0910 | 24.8203 | 9.5875 | 21.245 | 23.7718 | 18.9825 |
98
+ | 0.0007 | 45.0 | 37620 | 0.0914 | 25.1212 | 10.1024 | 21.6215 | 24.1061 | 18.9771 |
99
+ | 0.0006 | 46.0 | 38456 | 0.0914 | 25.1636 | 9.8127 | 21.5343 | 24.13 | 18.9475 |
100
+ | 0.0006 | 47.0 | 39292 | 0.0915 | 24.866 | 9.8427 | 21.3531 | 23.8643 | 18.9394 |
101
+ | 0.0006 | 48.0 | 40128 | 0.0916 | 25.064 | 10.049 | 21.5198 | 24.1158 | 18.9731 |
102
+ | 0.0005 | 49.0 | 40964 | 0.0923 | 24.8424 | 9.9718 | 21.3263 | 23.9031 | 18.9933 |
103
+ | 0.0005 | 50.0 | 41800 | 0.0923 | 25.0369 | 10.1572 | 21.5244 | 24.0775 | 18.9946 |
104
 
105
 
106
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b189ebcab14f5cc4166fc42a519fd883fd49274d8e85b4ae71a1fdaa58d1f86
3
  size 958009136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0bc740506bcb22a1bb7e5f72070b7ee8ee6aab692e13c32f35fc13070e77ab94
3
  size 958009136