hieungo1410 commited on
Commit
26f828e
1 Parent(s): dde9065

End of training

Browse files
Files changed (3) hide show
  1. README.md +58 -18
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -15,13 +15,13 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [VietAI/vit5-base](https://huggingface.co/VietAI/vit5-base) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.0439
19
- - Score: 42.2142
20
- - Counts: [2084, 1925, 1770, 1616]
21
- - Totals: [2111, 1964, 1817, 1670]
22
- - Precisions: [98.72098531501658, 98.0142566191446, 97.41331865712714, 96.76646706586827]
23
- - Bp: 0.4320
24
- - Sys Len: 2111
25
  - Ref Len: 3883
26
 
27
  ## Model description
@@ -47,22 +47,62 @@ The following hyperparameters were used during training:
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
- - num_epochs: 10
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Score | Counts | Totals | Precisions | Bp | Sys Len | Ref Len |
55
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:------------------------:|:------------------------:|:----------------------------------------------------------------------------:|:------:|:-------:|:-------:|
56
- | No log | 1.0 | 71 | 0.3049 | 34.8561 | [1920, 1622, 1384, 1161] | [2160, 2013, 1866, 1719] | [88.88888888888889, 80.57625434674615, 74.16934619506966, 67.53926701570681] | 0.4504 | 2160 | 3883 |
57
- | No log | 2.0 | 142 | 0.2064 | 37.7340 | [1984, 1738, 1528, 1321] | [2150, 2003, 1856, 1709] | [92.27906976744185, 86.76984523215177, 82.32758620689656, 77.29666471620831] | 0.4466 | 2150 | 3883 |
58
- | No log | 3.0 | 213 | 0.1438 | 39.7849 | [2029, 1813, 1629, 1452] | [2141, 1994, 1847, 1700] | [94.76879962634283, 90.92276830491474, 88.19707634001082, 85.41176470588235] | 0.4432 | 2141 | 3883 |
59
- | No log | 4.0 | 284 | 0.1157 | 40.8740 | [2050, 1864, 1693, 1524] | [2128, 1981, 1834, 1687] | [96.33458646616542, 94.09389197375063, 92.31188658669575, 90.33787788974512] | 0.4384 | 2128 | 3883 |
60
- | No log | 5.0 | 355 | 0.0860 | 41.2231 | [2066, 1889, 1722, 1559] | [2108, 1961, 1814, 1667] | [98.00759013282732, 96.32840387557368, 94.92833517089305, 93.52129574085183] | 0.4308 | 2108 | 3883 |
61
- | No log | 6.0 | 426 | 0.0727 | 41.5588 | [2068, 1897, 1738, 1583] | [2110, 1963, 1816, 1669] | [98.00947867298578, 96.63779928680592, 95.70484581497797, 94.84721390053924] | 0.4316 | 2110 | 3883 |
62
- | No log | 7.0 | 497 | 0.0564 | 41.9034 | [2076, 1914, 1759, 1607] | [2105, 1958, 1811, 1664] | [98.62232779097387, 97.75280898876404, 97.12865819988956, 96.57451923076923] | 0.4297 | 2105 | 3883 |
63
- | 0.3736 | 8.0 | 568 | 0.0497 | 42.1238 | [2080, 1919, 1763, 1608] | [2115, 1968, 1821, 1674] | [98.3451536643026, 97.51016260162602, 96.81493684788578, 96.05734767025089] | 0.4335 | 2115 | 3883 |
64
- | 0.3736 | 9.0 | 639 | 0.0449 | 42.2673 | [2084, 1927, 1774, 1621] | [2110, 1963, 1816, 1669] | [98.76777251184834, 98.16607233825776, 97.68722466960352, 97.12402636309167] | 0.4316 | 2110 | 3883 |
65
- | 0.3736 | 10.0 | 710 | 0.0439 | 42.2142 | [2084, 1925, 1770, 1616] | [2111, 1964, 1817, 1670] | [98.72098531501658, 98.0142566191446, 97.41331865712714, 96.76646706586827] | 0.4320 | 2111 | 3883 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
 
67
 
68
  ### Framework versions
 
15
 
16
  This model is a fine-tuned version of [VietAI/vit5-base](https://huggingface.co/VietAI/vit5-base) on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.0000
19
+ - Score: 42.9309
20
+ - Counts: [2102, 1955, 1808, 1661]
21
+ - Totals: [2107, 1960, 1813, 1666]
22
+ - Precisions: [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079]
23
+ - Bp: 0.4305
24
+ - Sys Len: 2107
25
  - Ref Len: 3883
26
 
27
  ## Model description
 
47
  - seed: 42
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - num_epochs: 50
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Score | Counts | Totals | Precisions | Bp | Sys Len | Ref Len |
55
  |:-------------:|:-----:|:----:|:---------------:|:-------:|:------------------------:|:------------------------:|:----------------------------------------------------------------------------:|:------:|:-------:|:-------:|
56
+ | No log | 1.0 | 71 | 0.3014 | 34.7852 | [1924, 1625, 1376, 1148] | [2164, 2017, 1870, 1723] | [88.909426987061, 80.56519583539911, 73.58288770053476, 66.62797446314568] | 0.4519 | 2164 | 3883 |
57
+ | No log | 2.0 | 142 | 0.1988 | 37.2103 | [1971, 1718, 1504, 1294] | [2149, 2002, 1855, 1708] | [91.71707771056305, 85.81418581418582, 81.07816711590297, 75.76112412177986] | 0.4462 | 2149 | 3883 |
58
+ | No log | 3.0 | 213 | 0.1335 | 39.9375 | [2031, 1820, 1638, 1463] | [2139, 1992, 1845, 1698] | [94.95091164095372, 91.36546184738955, 88.78048780487805, 86.16018845700825] | 0.4425 | 2139 | 3883 |
59
+ | No log | 4.0 | 284 | 0.1097 | 40.6507 | [2044, 1854, 1679, 1508] | [2132, 1985, 1838, 1691] | [95.87242026266416, 93.40050377833754, 91.34929270946681, 89.17800118273212] | 0.4399 | 2132 | 3883 |
60
+ | No log | 5.0 | 355 | 0.0707 | 40.8459 | [2053, 1879, 1715, 1554] | [2095, 1948, 1801, 1654] | [97.99522673031026, 96.45790554414785, 95.22487506940588, 93.9540507859734] | 0.4259 | 2095 | 3883 |
61
+ | No log | 6.0 | 426 | 0.0596 | 41.4723 | [2064, 1898, 1740, 1581] | [2105, 1958, 1811, 1664] | [98.05225653206651, 96.93564862104188, 96.07951408061844, 95.01201923076923] | 0.4297 | 2105 | 3883 |
62
+ | No log | 7.0 | 497 | 0.0457 | 41.8025 | [2077, 1916, 1758, 1601] | [2100, 1953, 1806, 1659] | [98.9047619047619, 98.10547875064005, 97.3421926910299, 96.50391802290537] | 0.4278 | 2100 | 3883 |
63
+ | 0.3603 | 8.0 | 568 | 0.0400 | 42.5405 | [2094, 1939, 1783, 1628] | [2114, 1967, 1820, 1673] | [99.05392620624409, 98.57651245551601, 97.96703296703296, 97.31022115959354] | 0.4331 | 2114 | 3883 |
64
+ | 0.3603 | 9.0 | 639 | 0.0345 | 42.6079 | [2096, 1940, 1783, 1625] | [2119, 1972, 1825, 1678] | [98.91458235016518, 98.37728194726166, 97.6986301369863, 96.84147794994041] | 0.4350 | 2119 | 3883 |
65
+ | 0.3603 | 10.0 | 710 | 0.0258 | 42.5911 | [2094, 1939, 1783, 1627] | [2118, 1971, 1824, 1677] | [98.86685552407933, 98.37645865043125, 97.75219298245614, 97.01848539057842] | 0.4346 | 2118 | 3883 |
66
+ | 0.3603 | 11.0 | 781 | 0.0237 | 42.7680 | [2097, 1947, 1798, 1649] | [2110, 1963, 1816, 1669] | [99.38388625592417, 99.18492103922567, 99.00881057268722, 98.8016776512882] | 0.4316 | 2110 | 3883 |
67
+ | 0.3603 | 12.0 | 852 | 0.0189 | 42.7976 | [2100, 1948, 1795, 1642] | [2115, 1968, 1821, 1674] | [99.29078014184397, 98.98373983739837, 98.5722130697419, 98.08841099163679] | 0.4335 | 2115 | 3883 |
68
+ | 0.3603 | 13.0 | 923 | 0.0131 | 42.9215 | [2101, 1952, 1803, 1654] | [2113, 1966, 1819, 1672] | [99.43208707998107, 99.28789420142421, 99.12039582188015, 98.92344497607655] | 0.4327 | 2113 | 3883 |
69
+ | 0.3603 | 14.0 | 994 | 0.0092 | 42.8090 | [2098, 1949, 1800, 1651] | [2110, 1963, 1816, 1669] | [99.43127962085308, 99.28680590932247, 99.11894273127753, 98.92150988615937] | 0.4316 | 2110 | 3883 |
70
+ | 0.0675 | 15.0 | 1065 | 0.0092 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
71
+ | 0.0675 | 16.0 | 1136 | 0.0078 | 42.8571 | [2103, 1954, 1803, 1652] | [2108, 1961, 1814, 1667] | [99.76280834914611, 99.64303926568077, 99.393605292172, 99.1001799640072] | 0.4308 | 2108 | 3883 |
72
+ | 0.0675 | 17.0 | 1207 | 0.0085 | 42.9060 | [2101, 1953, 1803, 1653] | [2112, 1965, 1818, 1671] | [99.47916666666667, 99.38931297709924, 99.17491749174917, 98.92280071813285] | 0.4323 | 2112 | 3883 |
73
+ | 0.0675 | 18.0 | 1278 | 0.0045 | 42.7799 | [2098, 1949, 1800, 1651] | [2108, 1961, 1814, 1667] | [99.52561669829223, 99.38806731259561, 99.22822491730982, 99.04019196160768] | 0.4308 | 2108 | 3883 |
74
+ | 0.0675 | 19.0 | 1349 | 0.0058 | 42.8931 | [2102, 1955, 1806, 1657] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.61389961389962, 99.45978391356543] | 0.4305 | 2107 | 3883 |
75
+ | 0.0675 | 20.0 | 1420 | 0.0052 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
76
+ | 0.0675 | 21.0 | 1491 | 0.0052 | 42.9512 | [2105, 1956, 1806, 1656] | [2110, 1963, 1816, 1669] | [99.76303317535545, 99.64340295466123, 99.44933920704845, 99.22109047333733] | 0.4316 | 2110 | 3883 |
77
+ | 0.0326 | 22.0 | 1562 | 0.0067 | 42.7869 | [2097, 1947, 1799, 1651] | [2110, 1963, 1816, 1669] | [99.38388625592417, 99.18492103922567, 99.06387665198238, 98.92150988615937] | 0.4316 | 2110 | 3883 |
78
+ | 0.0326 | 23.0 | 1633 | 0.0016 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
79
+ | 0.0326 | 24.0 | 1704 | 0.0044 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
80
+ | 0.0326 | 25.0 | 1775 | 0.0023 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
81
+ | 0.0326 | 26.0 | 1846 | 0.0010 | 42.9496 | [2103, 1956, 1808, 1660] | [2108, 1961, 1814, 1667] | [99.76280834914611, 99.74502804691484, 99.66923925027564, 99.58008398320337] | 0.4308 | 2108 | 3883 |
82
+ | 0.0326 | 27.0 | 1917 | 0.0011 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
83
+ | 0.0326 | 28.0 | 1988 | 0.0019 | 42.9317 | [2103, 1955, 1807, 1659] | [2108, 1961, 1814, 1667] | [99.76280834914611, 99.6940336562978, 99.6141124586549, 99.52009598080384] | 0.4308 | 2108 | 3883 |
84
+ | 0.0169 | 29.0 | 2059 | 0.0020 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
85
+ | 0.0169 | 30.0 | 2130 | 0.0015 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
86
+ | 0.0169 | 31.0 | 2201 | 0.0016 | 42.9120 | [2102, 1955, 1807, 1659] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.66905681191396, 99.5798319327731] | 0.4305 | 2107 | 3883 |
87
+ | 0.0169 | 32.0 | 2272 | 0.0012 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
88
+ | 0.0169 | 33.0 | 2343 | 0.0015 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
89
+ | 0.0169 | 34.0 | 2414 | 0.0022 | 42.8573 | [2102, 1953, 1804, 1655] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.64285714285714, 99.50358521787093, 99.33973589435774] | 0.4305 | 2107 | 3883 |
90
+ | 0.0169 | 35.0 | 2485 | 0.0005 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
91
+ | 0.0108 | 36.0 | 2556 | 0.0006 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
92
+ | 0.0108 | 37.0 | 2627 | 0.0009 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
93
+ | 0.0108 | 38.0 | 2698 | 0.0013 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
94
+ | 0.0108 | 39.0 | 2769 | 0.0010 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
95
+ | 0.0108 | 40.0 | 2840 | 0.0010 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
96
+ | 0.0108 | 41.0 | 2911 | 0.0002 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
97
+ | 0.0108 | 42.0 | 2982 | 0.0006 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
98
+ | 0.0069 | 43.0 | 3053 | 0.0002 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
99
+ | 0.0069 | 44.0 | 3124 | 0.0000 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
100
+ | 0.0069 | 45.0 | 3195 | 0.0000 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
101
+ | 0.0069 | 46.0 | 3266 | 0.0001 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
102
+ | 0.0069 | 47.0 | 3337 | 0.0000 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
103
+ | 0.0069 | 48.0 | 3408 | 0.0000 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
104
+ | 0.0069 | 49.0 | 3479 | 0.0000 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
105
+ | 0.0049 | 50.0 | 3550 | 0.0000 | 42.9309 | [2102, 1955, 1808, 1661] | [2107, 1960, 1813, 1666] | [99.76269577598481, 99.74489795918367, 99.72421400992829, 99.69987995198079] | 0.4305 | 2107 | 3883 |
106
 
107
 
108
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb520f01603de33e704d0f38a31595db85b20624a22a7031e7e7bea9f6754897
3
  size 903834408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:04af2cdd1cf899e43e86098e660641e25057aeff6af5932618c100cd53f7fc8e
3
  size 903834408
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4cf472288ddde8c53dad197634f9e45c5451c1931435fff4c71dfd05daaf79c1
3
  size 4411
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b77a6c0746fd7b09539bccd8dca42ff4e02ac74147310653f40bf0dd670dc8d6
3
  size 4411