Upload 11 files
Browse files- adapter_model.safetensors +1 -1
- optimizer.pt +1 -1
- scheduler.pt +1 -1
- trainer_state.json +174 -174
- training_args.bin +1 -1
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3416264
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:65ebc8867eb68b43b53f8b168d233d581038889786395a1deeb0793e650370fc
|
3 |
size 3416264
|
optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1807738
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92b7686e8135c430264a435243d5d21ee10202f63beca9991fc2970ea91dd57f
|
3 |
size 1807738
|
scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:22d3783e7365b2512df7ed24791ee332b6535a1eea8b7980b6eed32ed57b4061
|
3 |
size 1064
|
trainer_state.json
CHANGED
@@ -10,242 +10,242 @@
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.11851851851851852,
|
13 |
-
"grad_norm": 3.
|
14 |
-
"learning_rate":
|
15 |
-
"logits/chosen": 0.
|
16 |
-
"logits/rejected": 0.
|
17 |
-
"logps/chosen": -
|
18 |
-
"logps/rejected": -
|
19 |
-
"loss": 0.
|
20 |
-
"rewards/accuracies": 0.
|
21 |
-
"rewards/chosen": -0.
|
22 |
-
"rewards/margins": 0.
|
23 |
-
"rewards/rejected": -0.
|
24 |
"step": 4
|
25 |
},
|
26 |
{
|
27 |
"epoch": 0.23703703703703705,
|
28 |
-
"grad_norm": 4.
|
29 |
-
"learning_rate":
|
30 |
-
"logits/chosen": 0.
|
31 |
-
"logits/rejected": 0.
|
32 |
-
"logps/chosen": -
|
33 |
-
"logps/rejected": -
|
34 |
-
"loss": 0.
|
35 |
"rewards/accuracies": 0.71875,
|
36 |
-
"rewards/chosen": -0.
|
37 |
-
"rewards/margins": 0.
|
38 |
-
"rewards/rejected": -0.
|
39 |
"step": 8
|
40 |
},
|
41 |
{
|
42 |
"epoch": 0.35555555555555557,
|
43 |
-
"grad_norm":
|
44 |
-
"learning_rate":
|
45 |
-
"logits/chosen": 0.
|
46 |
-
"logits/rejected": 0.
|
47 |
-
"logps/chosen": -
|
48 |
-
"logps/rejected": -
|
49 |
-
"loss": 0.
|
50 |
-
"rewards/accuracies": 0.
|
51 |
-
"rewards/chosen": -0.
|
52 |
-
"rewards/margins": 0.
|
53 |
-
"rewards/rejected": -0.
|
54 |
"step": 12
|
55 |
},
|
56 |
{
|
57 |
"epoch": 0.4740740740740741,
|
58 |
-
"grad_norm":
|
59 |
-
"learning_rate":
|
60 |
-
"logits/chosen": 0.
|
61 |
-
"logits/rejected": 0.
|
62 |
-
"logps/chosen": -
|
63 |
-
"logps/rejected": -
|
64 |
-
"loss": 0.
|
65 |
"rewards/accuracies": 0.84375,
|
66 |
-
"rewards/chosen": -0.
|
67 |
-
"rewards/margins":
|
68 |
-
"rewards/rejected": -
|
69 |
"step": 16
|
70 |
},
|
71 |
{
|
72 |
"epoch": 0.5925925925925926,
|
73 |
-
"grad_norm":
|
74 |
-
"learning_rate":
|
75 |
-
"logits/chosen": 0.
|
76 |
-
"logits/rejected":
|
77 |
-
"logps/chosen": -
|
78 |
-
"logps/rejected": -
|
79 |
-
"loss": 0.
|
80 |
-
"rewards/accuracies": 0.
|
81 |
-
"rewards/chosen": -
|
82 |
-
"rewards/margins":
|
83 |
-
"rewards/rejected": -
|
84 |
"step": 20
|
85 |
},
|
86 |
{
|
87 |
"epoch": 0.7111111111111111,
|
88 |
-
"grad_norm":
|
89 |
-
"learning_rate":
|
90 |
-
"logits/chosen": 0.
|
91 |
-
"logits/rejected":
|
92 |
-
"logps/chosen": -
|
93 |
-
"logps/rejected": -
|
94 |
-
"loss": 0.
|
95 |
-
"rewards/accuracies": 0.
|
96 |
-
"rewards/chosen": -
|
97 |
-
"rewards/margins":
|
98 |
-
"rewards/rejected": -
|
99 |
"step": 24
|
100 |
},
|
101 |
{
|
102 |
"epoch": 0.8296296296296296,
|
103 |
-
"grad_norm":
|
104 |
-
"learning_rate":
|
105 |
-
"logits/chosen": 0.
|
106 |
-
"logits/rejected":
|
107 |
-
"logps/chosen": -
|
108 |
-
"logps/rejected": -
|
109 |
-
"loss": 0.
|
110 |
-
"rewards/accuracies": 0.
|
111 |
-
"rewards/chosen": -
|
112 |
-
"rewards/margins":
|
113 |
-
"rewards/rejected": -
|
114 |
"step": 28
|
115 |
},
|
116 |
{
|
117 |
"epoch": 0.9481481481481482,
|
118 |
-
"grad_norm":
|
119 |
-
"learning_rate":
|
120 |
-
"logits/chosen": 0.
|
121 |
-
"logits/rejected":
|
122 |
-
"logps/chosen": -
|
123 |
-
"logps/rejected": -
|
124 |
-
"loss": 0.
|
125 |
-
"rewards/accuracies": 0.
|
126 |
-
"rewards/chosen": -
|
127 |
-
"rewards/margins":
|
128 |
-
"rewards/rejected": -
|
129 |
"step": 32
|
130 |
},
|
131 |
{
|
132 |
"epoch": 1.0666666666666667,
|
133 |
-
"grad_norm":
|
134 |
-
"learning_rate":
|
135 |
-
"logits/chosen":
|
136 |
-
"logits/rejected":
|
137 |
-
"logps/chosen": -
|
138 |
-
"logps/rejected": -
|
139 |
-
"loss": 0.
|
140 |
-
"rewards/accuracies": 0.
|
141 |
-
"rewards/chosen": -
|
142 |
-
"rewards/margins":
|
143 |
-
"rewards/rejected": -
|
144 |
"step": 36
|
145 |
},
|
146 |
{
|
147 |
"epoch": 1.1851851851851851,
|
148 |
-
"grad_norm":
|
149 |
-
"learning_rate":
|
150 |
-
"logits/chosen":
|
151 |
-
"logits/rejected":
|
152 |
-
"logps/chosen": -
|
153 |
-
"logps/rejected": -
|
154 |
-
"loss": 0.
|
155 |
-
"rewards/accuracies": 0.
|
156 |
-
"rewards/chosen": -
|
157 |
-
"rewards/margins":
|
158 |
-
"rewards/rejected": -
|
159 |
"step": 40
|
160 |
},
|
161 |
{
|
162 |
"epoch": 1.3037037037037038,
|
163 |
-
"grad_norm":
|
164 |
-
"learning_rate":
|
165 |
-
"logits/chosen": 0.
|
166 |
-
"logits/rejected":
|
167 |
-
"logps/chosen": -
|
168 |
-
"logps/rejected": -
|
169 |
-
"loss": 0.
|
170 |
-
"rewards/accuracies": 0.
|
171 |
-
"rewards/chosen": -
|
172 |
-
"rewards/margins":
|
173 |
-
"rewards/rejected": -
|
174 |
"step": 44
|
175 |
},
|
176 |
{
|
177 |
"epoch": 1.4222222222222223,
|
178 |
-
"grad_norm":
|
179 |
-
"learning_rate":
|
180 |
-
"logits/chosen": 0.
|
181 |
-
"logits/rejected":
|
182 |
-
"logps/chosen": -
|
183 |
-
"logps/rejected": -
|
184 |
-
"loss": 0.
|
185 |
-
"rewards/accuracies": 0.
|
186 |
-
"rewards/chosen": -
|
187 |
-
"rewards/margins":
|
188 |
-
"rewards/rejected": -
|
189 |
"step": 48
|
190 |
},
|
191 |
{
|
192 |
"epoch": 1.5407407407407407,
|
193 |
-
"grad_norm":
|
194 |
-
"learning_rate":
|
195 |
-
"logits/chosen": 0.
|
196 |
-
"logits/rejected": -0.
|
197 |
-
"logps/chosen": -148.
|
198 |
-
"logps/rejected": -
|
199 |
-
"loss": 0.
|
200 |
-
"rewards/accuracies": 0.
|
201 |
-
"rewards/chosen": -
|
202 |
-
"rewards/margins":
|
203 |
-
"rewards/rejected": -
|
204 |
"step": 52
|
205 |
},
|
206 |
{
|
207 |
"epoch": 1.6592592592592592,
|
208 |
-
"grad_norm":
|
209 |
-
"learning_rate":
|
210 |
-
"logits/chosen": 0.
|
211 |
-
"logits/rejected":
|
212 |
-
"logps/chosen": -
|
213 |
-
"logps/rejected": -
|
214 |
-
"loss": 0.
|
215 |
-
"rewards/accuracies": 0.
|
216 |
-
"rewards/chosen": -
|
217 |
-
"rewards/margins":
|
218 |
-
"rewards/rejected": -
|
219 |
"step": 56
|
220 |
},
|
221 |
{
|
222 |
"epoch": 1.7777777777777777,
|
223 |
-
"grad_norm":
|
224 |
-
"learning_rate":
|
225 |
-
"logits/chosen": 0.
|
226 |
-
"logits/rejected":
|
227 |
-
"logps/chosen": -
|
228 |
-
"logps/rejected": -
|
229 |
-
"loss": 0.
|
230 |
-
"rewards/accuracies": 0.
|
231 |
-
"rewards/chosen": -
|
232 |
-
"rewards/margins":
|
233 |
-
"rewards/rejected": -
|
234 |
"step": 60
|
235 |
},
|
236 |
{
|
237 |
"epoch": 1.8962962962962964,
|
238 |
-
"grad_norm":
|
239 |
-
"learning_rate":
|
240 |
-
"logits/chosen": 0.
|
241 |
-
"logits/rejected":
|
242 |
-
"logps/chosen": -
|
243 |
-
"logps/rejected": -
|
244 |
-
"loss": 0.
|
245 |
-
"rewards/accuracies": 0.
|
246 |
-
"rewards/chosen": -
|
247 |
-
"rewards/margins":
|
248 |
-
"rewards/rejected": -
|
249 |
"step": 64
|
250 |
}
|
251 |
],
|
|
|
10 |
"log_history": [
|
11 |
{
|
12 |
"epoch": 0.11851851851851852,
|
13 |
+
"grad_norm": 3.8725790977478027,
|
14 |
+
"learning_rate": 5.714285714285714e-05,
|
15 |
+
"logits/chosen": 0.5535547137260437,
|
16 |
+
"logits/rejected": 0.6790706515312195,
|
17 |
+
"logps/chosen": -154.02122497558594,
|
18 |
+
"logps/rejected": -190.33836364746094,
|
19 |
+
"loss": 0.6916,
|
20 |
+
"rewards/accuracies": 0.28125,
|
21 |
+
"rewards/chosen": -0.00438268156722188,
|
22 |
+
"rewards/margins": 0.0031897304579615593,
|
23 |
+
"rewards/rejected": -0.0075724124908447266,
|
24 |
"step": 4
|
25 |
},
|
26 |
{
|
27 |
"epoch": 0.23703703703703705,
|
28 |
+
"grad_norm": 4.261082649230957,
|
29 |
+
"learning_rate": 9.830508474576272e-05,
|
30 |
+
"logits/chosen": 0.8175854086875916,
|
31 |
+
"logits/rejected": 0.7639511823654175,
|
32 |
+
"logps/chosen": -176.618408203125,
|
33 |
+
"logps/rejected": -232.55316162109375,
|
34 |
+
"loss": 0.6661,
|
35 |
"rewards/accuracies": 0.71875,
|
36 |
+
"rewards/chosen": -0.024643946439027786,
|
37 |
+
"rewards/margins": 0.05728544294834137,
|
38 |
+
"rewards/rejected": -0.08192938566207886,
|
39 |
"step": 8
|
40 |
},
|
41 |
{
|
42 |
"epoch": 0.35555555555555557,
|
43 |
+
"grad_norm": 3.7274258136749268,
|
44 |
+
"learning_rate": 9.152542372881357e-05,
|
45 |
+
"logits/chosen": 0.5340245366096497,
|
46 |
+
"logits/rejected": 0.748319685459137,
|
47 |
+
"logps/chosen": -137.61412048339844,
|
48 |
+
"logps/rejected": -184.79991149902344,
|
49 |
+
"loss": 0.6328,
|
50 |
+
"rewards/accuracies": 0.71875,
|
51 |
+
"rewards/chosen": -0.10764569044113159,
|
52 |
+
"rewards/margins": 0.14540965855121613,
|
53 |
+
"rewards/rejected": -0.2530553638935089,
|
54 |
"step": 12
|
55 |
},
|
56 |
{
|
57 |
"epoch": 0.4740740740740741,
|
58 |
+
"grad_norm": 4.1018290519714355,
|
59 |
+
"learning_rate": 8.474576271186441e-05,
|
60 |
+
"logits/chosen": 0.43400460481643677,
|
61 |
+
"logits/rejected": 0.4534505605697632,
|
62 |
+
"logps/chosen": -113.1705551147461,
|
63 |
+
"logps/rejected": -201.7666778564453,
|
64 |
+
"loss": 0.5399,
|
65 |
"rewards/accuracies": 0.84375,
|
66 |
+
"rewards/chosen": -0.11035704612731934,
|
67 |
+
"rewards/margins": 0.3914012014865875,
|
68 |
+
"rewards/rejected": -0.5017582178115845,
|
69 |
"step": 16
|
70 |
},
|
71 |
{
|
72 |
"epoch": 0.5925925925925926,
|
73 |
+
"grad_norm": 7.585125923156738,
|
74 |
+
"learning_rate": 7.796610169491526e-05,
|
75 |
+
"logits/chosen": 0.3885829448699951,
|
76 |
+
"logits/rejected": 0.4270927906036377,
|
77 |
+
"logps/chosen": -162.44253540039062,
|
78 |
+
"logps/rejected": -242.7151336669922,
|
79 |
+
"loss": 0.5997,
|
80 |
+
"rewards/accuracies": 0.65625,
|
81 |
+
"rewards/chosen": -0.5063179731369019,
|
82 |
+
"rewards/margins": 0.40314924716949463,
|
83 |
+
"rewards/rejected": -0.9094672799110413,
|
84 |
"step": 20
|
85 |
},
|
86 |
{
|
87 |
"epoch": 0.7111111111111111,
|
88 |
+
"grad_norm": 4.802177429199219,
|
89 |
+
"learning_rate": 7.11864406779661e-05,
|
90 |
+
"logits/chosen": 0.39339229464530945,
|
91 |
+
"logits/rejected": 0.1176382452249527,
|
92 |
+
"logps/chosen": -157.732177734375,
|
93 |
+
"logps/rejected": -208.583251953125,
|
94 |
+
"loss": 0.4315,
|
95 |
+
"rewards/accuracies": 0.8125,
|
96 |
+
"rewards/chosen": -0.5111438632011414,
|
97 |
+
"rewards/margins": 0.9935877323150635,
|
98 |
+
"rewards/rejected": -1.5047316551208496,
|
99 |
"step": 24
|
100 |
},
|
101 |
{
|
102 |
"epoch": 0.8296296296296296,
|
103 |
+
"grad_norm": 8.156561851501465,
|
104 |
+
"learning_rate": 6.440677966101695e-05,
|
105 |
+
"logits/chosen": 0.011842329055070877,
|
106 |
+
"logits/rejected": 0.35012856125831604,
|
107 |
+
"logps/chosen": -158.82598876953125,
|
108 |
+
"logps/rejected": -232.102783203125,
|
109 |
+
"loss": 0.6494,
|
110 |
+
"rewards/accuracies": 0.6875,
|
111 |
+
"rewards/chosen": -0.8290945291519165,
|
112 |
+
"rewards/margins": 0.6025375723838806,
|
113 |
+
"rewards/rejected": -1.4316319227218628,
|
114 |
"step": 28
|
115 |
},
|
116 |
{
|
117 |
"epoch": 0.9481481481481482,
|
118 |
+
"grad_norm": 6.632073402404785,
|
119 |
+
"learning_rate": 5.76271186440678e-05,
|
120 |
+
"logits/chosen": -0.034538522362709045,
|
121 |
+
"logits/rejected": 0.06912083178758621,
|
122 |
+
"logps/chosen": -163.13931274414062,
|
123 |
+
"logps/rejected": -159.48757934570312,
|
124 |
+
"loss": 0.6065,
|
125 |
+
"rewards/accuracies": 0.78125,
|
126 |
+
"rewards/chosen": -1.0138874053955078,
|
127 |
+
"rewards/margins": 0.7364886403083801,
|
128 |
+
"rewards/rejected": -1.7503761053085327,
|
129 |
"step": 32
|
130 |
},
|
131 |
{
|
132 |
"epoch": 1.0666666666666667,
|
133 |
+
"grad_norm": 4.534868240356445,
|
134 |
+
"learning_rate": 5.0847457627118643e-05,
|
135 |
+
"logits/chosen": 0.30354946851730347,
|
136 |
+
"logits/rejected": 0.13392731547355652,
|
137 |
+
"logps/chosen": -154.13514709472656,
|
138 |
+
"logps/rejected": -247.21238708496094,
|
139 |
+
"loss": 0.4295,
|
140 |
+
"rewards/accuracies": 0.78125,
|
141 |
+
"rewards/chosen": -0.7905898094177246,
|
142 |
+
"rewards/margins": 1.3606148958206177,
|
143 |
+
"rewards/rejected": -2.1512045860290527,
|
144 |
"step": 36
|
145 |
},
|
146 |
{
|
147 |
"epoch": 1.1851851851851851,
|
148 |
+
"grad_norm": 8.053945541381836,
|
149 |
+
"learning_rate": 4.4067796610169495e-05,
|
150 |
+
"logits/chosen": 0.31644004583358765,
|
151 |
+
"logits/rejected": 0.34593597054481506,
|
152 |
+
"logps/chosen": -198.6423797607422,
|
153 |
+
"logps/rejected": -201.91390991210938,
|
154 |
+
"loss": 0.4743,
|
155 |
+
"rewards/accuracies": 0.8125,
|
156 |
+
"rewards/chosen": -0.8822494149208069,
|
157 |
+
"rewards/margins": 0.9695678949356079,
|
158 |
+
"rewards/rejected": -1.8518173694610596,
|
159 |
"step": 40
|
160 |
},
|
161 |
{
|
162 |
"epoch": 1.3037037037037038,
|
163 |
+
"grad_norm": 3.4382383823394775,
|
164 |
+
"learning_rate": 3.728813559322034e-05,
|
165 |
+
"logits/chosen": 0.21577686071395874,
|
166 |
+
"logits/rejected": 0.19070810079574585,
|
167 |
+
"logps/chosen": -151.60324096679688,
|
168 |
+
"logps/rejected": -234.2410888671875,
|
169 |
+
"loss": 0.3084,
|
170 |
+
"rewards/accuracies": 0.9375,
|
171 |
+
"rewards/chosen": -0.48244649171829224,
|
172 |
+
"rewards/margins": 1.5388765335083008,
|
173 |
+
"rewards/rejected": -2.0213229656219482,
|
174 |
"step": 44
|
175 |
},
|
176 |
{
|
177 |
"epoch": 1.4222222222222223,
|
178 |
+
"grad_norm": 3.3008296489715576,
|
179 |
+
"learning_rate": 3.050847457627119e-05,
|
180 |
+
"logits/chosen": 0.20948533713817596,
|
181 |
+
"logits/rejected": 0.24637725949287415,
|
182 |
+
"logps/chosen": -140.76052856445312,
|
183 |
+
"logps/rejected": -215.05499267578125,
|
184 |
+
"loss": 0.2868,
|
185 |
+
"rewards/accuracies": 0.9375,
|
186 |
+
"rewards/chosen": -0.4717501699924469,
|
187 |
+
"rewards/margins": 1.771106243133545,
|
188 |
+
"rewards/rejected": -2.242856502532959,
|
189 |
"step": 48
|
190 |
},
|
191 |
{
|
192 |
"epoch": 1.5407407407407407,
|
193 |
+
"grad_norm": 11.502059936523438,
|
194 |
+
"learning_rate": 2.3728813559322036e-05,
|
195 |
+
"logits/chosen": 0.08824478089809418,
|
196 |
+
"logits/rejected": -0.027387090027332306,
|
197 |
+
"logps/chosen": -148.2930450439453,
|
198 |
+
"logps/rejected": -225.59078979492188,
|
199 |
+
"loss": 0.5723,
|
200 |
+
"rewards/accuracies": 0.6875,
|
201 |
+
"rewards/chosen": -1.0404824018478394,
|
202 |
+
"rewards/margins": 1.086774230003357,
|
203 |
+
"rewards/rejected": -2.1272566318511963,
|
204 |
"step": 52
|
205 |
},
|
206 |
{
|
207 |
"epoch": 1.6592592592592592,
|
208 |
+
"grad_norm": 4.403205394744873,
|
209 |
+
"learning_rate": 1.694915254237288e-05,
|
210 |
+
"logits/chosen": 0.2315160483121872,
|
211 |
+
"logits/rejected": 0.18600693345069885,
|
212 |
+
"logps/chosen": -134.02783203125,
|
213 |
+
"logps/rejected": -226.99459838867188,
|
214 |
+
"loss": 0.2737,
|
215 |
+
"rewards/accuracies": 0.875,
|
216 |
+
"rewards/chosen": -0.5924757719039917,
|
217 |
+
"rewards/margins": 1.9642682075500488,
|
218 |
+
"rewards/rejected": -2.55674409866333,
|
219 |
"step": 56
|
220 |
},
|
221 |
{
|
222 |
"epoch": 1.7777777777777777,
|
223 |
+
"grad_norm": 4.00092077255249,
|
224 |
+
"learning_rate": 1.016949152542373e-05,
|
225 |
+
"logits/chosen": 0.10627911239862442,
|
226 |
+
"logits/rejected": 0.08208482712507248,
|
227 |
+
"logps/chosen": -172.69593811035156,
|
228 |
+
"logps/rejected": -211.60107421875,
|
229 |
+
"loss": 0.3421,
|
230 |
+
"rewards/accuracies": 0.875,
|
231 |
+
"rewards/chosen": -0.770930826663971,
|
232 |
+
"rewards/margins": 1.9104766845703125,
|
233 |
+
"rewards/rejected": -2.6814074516296387,
|
234 |
"step": 60
|
235 |
},
|
236 |
{
|
237 |
"epoch": 1.8962962962962964,
|
238 |
+
"grad_norm": 4.906458854675293,
|
239 |
+
"learning_rate": 3.3898305084745763e-06,
|
240 |
+
"logits/chosen": -0.028121717274188995,
|
241 |
+
"logits/rejected": 0.011860378086566925,
|
242 |
+
"logps/chosen": -121.95062255859375,
|
243 |
+
"logps/rejected": -200.142578125,
|
244 |
+
"loss": 0.3773,
|
245 |
+
"rewards/accuracies": 0.84375,
|
246 |
+
"rewards/chosen": -0.47138434648513794,
|
247 |
+
"rewards/margins": 1.7951738834381104,
|
248 |
+
"rewards/rejected": -2.2665581703186035,
|
249 |
"step": 64
|
250 |
}
|
251 |
],
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5112
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9a532c86fda2fd7158aecbe3b2707d774f9c356acd0b2658b92c3da108c9a58e
|
3 |
size 5112
|