alkatraz445 commited on
Commit
b78b249
·
verified ·
1 Parent(s): 5b1ef85

Upload folder using huggingface_hub

Browse files
checkpoint-35710/config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Real",
13
+ "1": "Fake"
14
+ },
15
+ "image_size": 224,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 3072,
18
+ "label2id": {
19
+ "Fake": 1,
20
+ "Real": 0
21
+ },
22
+ "layer_norm_eps": 1e-12,
23
+ "model_type": "vit",
24
+ "num_attention_heads": 12,
25
+ "num_channels": 3,
26
+ "num_hidden_layers": 12,
27
+ "patch_size": 16,
28
+ "problem_type": "single_label_classification",
29
+ "qkv_bias": true,
30
+ "torch_dtype": "float32",
31
+ "transformers_version": "4.46.0"
32
+ }
checkpoint-35710/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1fd12277336e8213bb338e1897cc234131944b7594acb3751f01821dadabab9
3
+ size 343223968
checkpoint-35710/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:22742cd7bab4c021f3d2566ef842d3f34e97d2f00f32b38b93ba79a4cc4bb95e
3
+ size 686568890
checkpoint-35710/preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
checkpoint-35710/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c966c8de452f8d7d4d60ed6c1960361bea0f03b86f9b775d412447b7f72cbe0
3
+ size 14244
checkpoint-35710/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c350ee2b771cabc13584f75deaea0b6208e30294d572bc65427cfb7ea10580a2
3
+ size 1064
checkpoint-35710/trainer_state.json ADDED
@@ -0,0 +1,630 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.11153655499219894,
3
+ "best_model_checkpoint": "deepfake_detection/checkpoint-35710",
4
+ "epoch": 10.0,
5
+ "eval_steps": 500,
6
+ "global_step": 35710,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.14001680201624195,
13
+ "grad_norm": 1.0528628826141357,
14
+ "learning_rate": 9.873808188446438e-07,
15
+ "loss": 0.6555,
16
+ "step": 500
17
+ },
18
+ {
19
+ "epoch": 0.2800336040324839,
20
+ "grad_norm": 1.7406058311462402,
21
+ "learning_rate": 9.733595064498036e-07,
22
+ "loss": 0.4817,
23
+ "step": 1000
24
+ },
25
+ {
26
+ "epoch": 0.42005040604872584,
27
+ "grad_norm": 2.337244987487793,
28
+ "learning_rate": 9.593381940549635e-07,
29
+ "loss": 0.3393,
30
+ "step": 1500
31
+ },
32
+ {
33
+ "epoch": 0.5600672080649678,
34
+ "grad_norm": 2.840176582336426,
35
+ "learning_rate": 9.453168816601234e-07,
36
+ "loss": 0.2739,
37
+ "step": 2000
38
+ },
39
+ {
40
+ "epoch": 0.7000840100812098,
41
+ "grad_norm": 2.7442541122436523,
42
+ "learning_rate": 9.312955692652832e-07,
43
+ "loss": 0.2335,
44
+ "step": 2500
45
+ },
46
+ {
47
+ "epoch": 0.8401008120974517,
48
+ "grad_norm": 4.048594951629639,
49
+ "learning_rate": 9.17274256870443e-07,
50
+ "loss": 0.2087,
51
+ "step": 3000
52
+ },
53
+ {
54
+ "epoch": 0.9801176141136937,
55
+ "grad_norm": 2.741589069366455,
56
+ "learning_rate": 9.032529444756028e-07,
57
+ "loss": 0.1953,
58
+ "step": 3500
59
+ },
60
+ {
61
+ "epoch": 1.0,
62
+ "eval_accuracy": 0.9271411877470096,
63
+ "eval_loss": 0.277784526348114,
64
+ "eval_model_preparation_time": 0.003,
65
+ "eval_runtime": 461.9798,
66
+ "eval_samples_per_second": 164.858,
67
+ "eval_steps_per_second": 20.609,
68
+ "step": 3571
69
+ },
70
+ {
71
+ "epoch": 1.1201344161299356,
72
+ "grad_norm": 2.69423246383667,
73
+ "learning_rate": 8.892316320807627e-07,
74
+ "loss": 0.1865,
75
+ "step": 4000
76
+ },
77
+ {
78
+ "epoch": 1.2601512181461776,
79
+ "grad_norm": 3.063028335571289,
80
+ "learning_rate": 8.752103196859225e-07,
81
+ "loss": 0.1789,
82
+ "step": 4500
83
+ },
84
+ {
85
+ "epoch": 1.4001680201624196,
86
+ "grad_norm": 2.7112696170806885,
87
+ "learning_rate": 8.611890072910825e-07,
88
+ "loss": 0.1587,
89
+ "step": 5000
90
+ },
91
+ {
92
+ "epoch": 1.5401848221786616,
93
+ "grad_norm": 1.334972858428955,
94
+ "learning_rate": 8.471676948962423e-07,
95
+ "loss": 0.1654,
96
+ "step": 5500
97
+ },
98
+ {
99
+ "epoch": 1.6802016241949034,
100
+ "grad_norm": 4.083142280578613,
101
+ "learning_rate": 8.331463825014021e-07,
102
+ "loss": 0.1472,
103
+ "step": 6000
104
+ },
105
+ {
106
+ "epoch": 1.8202184262111454,
107
+ "grad_norm": 3.0988452434539795,
108
+ "learning_rate": 8.191250701065619e-07,
109
+ "loss": 0.1454,
110
+ "step": 6500
111
+ },
112
+ {
113
+ "epoch": 1.9602352282273872,
114
+ "grad_norm": 4.7256011962890625,
115
+ "learning_rate": 8.051037577117218e-07,
116
+ "loss": 0.137,
117
+ "step": 7000
118
+ },
119
+ {
120
+ "epoch": 2.0,
121
+ "eval_accuracy": 0.9447748847835507,
122
+ "eval_loss": 0.19382010400295258,
123
+ "eval_model_preparation_time": 0.003,
124
+ "eval_runtime": 446.387,
125
+ "eval_samples_per_second": 170.617,
126
+ "eval_steps_per_second": 21.329,
127
+ "step": 7142
128
+ },
129
+ {
130
+ "epoch": 2.100252030243629,
131
+ "grad_norm": 2.3897340297698975,
132
+ "learning_rate": 7.910824453168816e-07,
133
+ "loss": 0.1272,
134
+ "step": 7500
135
+ },
136
+ {
137
+ "epoch": 2.240268832259871,
138
+ "grad_norm": 7.572381019592285,
139
+ "learning_rate": 7.770611329220415e-07,
140
+ "loss": 0.1339,
141
+ "step": 8000
142
+ },
143
+ {
144
+ "epoch": 2.380285634276113,
145
+ "grad_norm": 2.5557875633239746,
146
+ "learning_rate": 7.630398205272013e-07,
147
+ "loss": 0.1352,
148
+ "step": 8500
149
+ },
150
+ {
151
+ "epoch": 2.520302436292355,
152
+ "grad_norm": 2.867504119873047,
153
+ "learning_rate": 7.490185081323611e-07,
154
+ "loss": 0.1312,
155
+ "step": 9000
156
+ },
157
+ {
158
+ "epoch": 2.660319238308597,
159
+ "grad_norm": 5.774721622467041,
160
+ "learning_rate": 7.34997195737521e-07,
161
+ "loss": 0.1263,
162
+ "step": 9500
163
+ },
164
+ {
165
+ "epoch": 2.800336040324839,
166
+ "grad_norm": 1.3829355239868164,
167
+ "learning_rate": 7.209758833426809e-07,
168
+ "loss": 0.1222,
169
+ "step": 10000
170
+ },
171
+ {
172
+ "epoch": 2.940352842341081,
173
+ "grad_norm": 3.2408864498138428,
174
+ "learning_rate": 7.069545709478407e-07,
175
+ "loss": 0.1238,
176
+ "step": 10500
177
+ },
178
+ {
179
+ "epoch": 3.0,
180
+ "eval_accuracy": 0.9503420385761742,
181
+ "eval_loss": 0.16304399073123932,
182
+ "eval_model_preparation_time": 0.003,
183
+ "eval_runtime": 445.7116,
184
+ "eval_samples_per_second": 170.875,
185
+ "eval_steps_per_second": 21.361,
186
+ "step": 10713
187
+ },
188
+ {
189
+ "epoch": 3.0803696443573227,
190
+ "grad_norm": 5.649173259735107,
191
+ "learning_rate": 6.929332585530005e-07,
192
+ "loss": 0.1163,
193
+ "step": 11000
194
+ },
195
+ {
196
+ "epoch": 3.2203864463735647,
197
+ "grad_norm": 6.677682876586914,
198
+ "learning_rate": 6.789119461581604e-07,
199
+ "loss": 0.1151,
200
+ "step": 11500
201
+ },
202
+ {
203
+ "epoch": 3.3604032483898068,
204
+ "grad_norm": 6.577985763549805,
205
+ "learning_rate": 6.648906337633202e-07,
206
+ "loss": 0.116,
207
+ "step": 12000
208
+ },
209
+ {
210
+ "epoch": 3.5004200504060488,
211
+ "grad_norm": 4.521468162536621,
212
+ "learning_rate": 6.508693213684801e-07,
213
+ "loss": 0.1166,
214
+ "step": 12500
215
+ },
216
+ {
217
+ "epoch": 3.6404368524222908,
218
+ "grad_norm": 0.37339428067207336,
219
+ "learning_rate": 6.368480089736399e-07,
220
+ "loss": 0.1063,
221
+ "step": 13000
222
+ },
223
+ {
224
+ "epoch": 3.7804536544385328,
225
+ "grad_norm": 2.2542712688446045,
226
+ "learning_rate": 6.228266965787997e-07,
227
+ "loss": 0.108,
228
+ "step": 13500
229
+ },
230
+ {
231
+ "epoch": 3.9204704564547743,
232
+ "grad_norm": 11.904646873474121,
233
+ "learning_rate": 6.088053841839595e-07,
234
+ "loss": 0.1094,
235
+ "step": 14000
236
+ },
237
+ {
238
+ "epoch": 4.0,
239
+ "eval_accuracy": 0.9547406152755348,
240
+ "eval_loss": 0.14307229220867157,
241
+ "eval_model_preparation_time": 0.003,
242
+ "eval_runtime": 473.4476,
243
+ "eval_samples_per_second": 160.865,
244
+ "eval_steps_per_second": 20.11,
245
+ "step": 14284
246
+ },
247
+ {
248
+ "epoch": 4.060487258471016,
249
+ "grad_norm": 1.572487711906433,
250
+ "learning_rate": 5.947840717891194e-07,
251
+ "loss": 0.1105,
252
+ "step": 14500
253
+ },
254
+ {
255
+ "epoch": 4.200504060487258,
256
+ "grad_norm": 2.691279888153076,
257
+ "learning_rate": 5.807627593942794e-07,
258
+ "loss": 0.102,
259
+ "step": 15000
260
+ },
261
+ {
262
+ "epoch": 4.3405208625035,
263
+ "grad_norm": 3.9856207370758057,
264
+ "learning_rate": 5.667414469994392e-07,
265
+ "loss": 0.1032,
266
+ "step": 15500
267
+ },
268
+ {
269
+ "epoch": 4.480537664519742,
270
+ "grad_norm": 1.828147292137146,
271
+ "learning_rate": 5.52720134604599e-07,
272
+ "loss": 0.1029,
273
+ "step": 16000
274
+ },
275
+ {
276
+ "epoch": 4.620554466535984,
277
+ "grad_norm": 0.4480103850364685,
278
+ "learning_rate": 5.386988222097588e-07,
279
+ "loss": 0.1024,
280
+ "step": 16500
281
+ },
282
+ {
283
+ "epoch": 4.760571268552226,
284
+ "grad_norm": 9.204968452453613,
285
+ "learning_rate": 5.246775098149186e-07,
286
+ "loss": 0.0991,
287
+ "step": 17000
288
+ },
289
+ {
290
+ "epoch": 4.900588070568468,
291
+ "grad_norm": 7.890961170196533,
292
+ "learning_rate": 5.106561974200784e-07,
293
+ "loss": 0.1056,
294
+ "step": 17500
295
+ },
296
+ {
297
+ "epoch": 5.0,
298
+ "eval_accuracy": 0.9586927692651095,
299
+ "eval_loss": 0.12978993356227875,
300
+ "eval_model_preparation_time": 0.003,
301
+ "eval_runtime": 437.804,
302
+ "eval_samples_per_second": 173.961,
303
+ "eval_steps_per_second": 21.747,
304
+ "step": 17855
305
+ },
306
+ {
307
+ "epoch": 5.04060487258471,
308
+ "grad_norm": 2.9656715393066406,
309
+ "learning_rate": 4.966348850252384e-07,
310
+ "loss": 0.0981,
311
+ "step": 18000
312
+ },
313
+ {
314
+ "epoch": 5.180621674600952,
315
+ "grad_norm": 1.717795968055725,
316
+ "learning_rate": 4.826135726303982e-07,
317
+ "loss": 0.1049,
318
+ "step": 18500
319
+ },
320
+ {
321
+ "epoch": 5.320638476617194,
322
+ "grad_norm": 4.466497421264648,
323
+ "learning_rate": 4.6859226023555804e-07,
324
+ "loss": 0.1009,
325
+ "step": 19000
326
+ },
327
+ {
328
+ "epoch": 5.460655278633436,
329
+ "grad_norm": 2.382636547088623,
330
+ "learning_rate": 4.5457094784071786e-07,
331
+ "loss": 0.097,
332
+ "step": 19500
333
+ },
334
+ {
335
+ "epoch": 5.600672080649678,
336
+ "grad_norm": 3.537141799926758,
337
+ "learning_rate": 4.405496354458777e-07,
338
+ "loss": 0.0947,
339
+ "step": 20000
340
+ },
341
+ {
342
+ "epoch": 5.7406888826659195,
343
+ "grad_norm": 3.2444217205047607,
344
+ "learning_rate": 4.2652832305103755e-07,
345
+ "loss": 0.096,
346
+ "step": 20500
347
+ },
348
+ {
349
+ "epoch": 5.8807056846821615,
350
+ "grad_norm": 6.093824863433838,
351
+ "learning_rate": 4.125070106561974e-07,
352
+ "loss": 0.0962,
353
+ "step": 21000
354
+ },
355
+ {
356
+ "epoch": 6.0,
357
+ "eval_accuracy": 0.9606754112997465,
358
+ "eval_loss": 0.1219368726015091,
359
+ "eval_model_preparation_time": 0.003,
360
+ "eval_runtime": 461.6477,
361
+ "eval_samples_per_second": 164.976,
362
+ "eval_steps_per_second": 20.624,
363
+ "step": 21426
364
+ },
365
+ {
366
+ "epoch": 6.0207224866984035,
367
+ "grad_norm": 6.273184299468994,
368
+ "learning_rate": 3.9848569826135723e-07,
369
+ "loss": 0.1001,
370
+ "step": 21500
371
+ },
372
+ {
373
+ "epoch": 6.1607392887146455,
374
+ "grad_norm": 1.6552726030349731,
375
+ "learning_rate": 3.844643858665171e-07,
376
+ "loss": 0.0935,
377
+ "step": 22000
378
+ },
379
+ {
380
+ "epoch": 6.3007560907308875,
381
+ "grad_norm": 1.253029465675354,
382
+ "learning_rate": 3.704430734716769e-07,
383
+ "loss": 0.0944,
384
+ "step": 22500
385
+ },
386
+ {
387
+ "epoch": 6.4407728927471295,
388
+ "grad_norm": 6.506760120391846,
389
+ "learning_rate": 3.564217610768368e-07,
390
+ "loss": 0.092,
391
+ "step": 23000
392
+ },
393
+ {
394
+ "epoch": 6.5807896947633715,
395
+ "grad_norm": 6.743386268615723,
396
+ "learning_rate": 3.4240044868199666e-07,
397
+ "loss": 0.0955,
398
+ "step": 23500
399
+ },
400
+ {
401
+ "epoch": 6.7208064967796135,
402
+ "grad_norm": 7.667580604553223,
403
+ "learning_rate": 3.2837913628715647e-07,
404
+ "loss": 0.0923,
405
+ "step": 24000
406
+ },
407
+ {
408
+ "epoch": 6.8608232987958555,
409
+ "grad_norm": 3.372116804122925,
410
+ "learning_rate": 3.143578238923163e-07,
411
+ "loss": 0.0992,
412
+ "step": 24500
413
+ },
414
+ {
415
+ "epoch": 7.0,
416
+ "eval_accuracy": 0.9620934599073017,
417
+ "eval_loss": 0.11666399985551834,
418
+ "eval_model_preparation_time": 0.003,
419
+ "eval_runtime": 447.5043,
420
+ "eval_samples_per_second": 170.191,
421
+ "eval_steps_per_second": 21.276,
422
+ "step": 24997
423
+ },
424
+ {
425
+ "epoch": 7.0008401008120975,
426
+ "grad_norm": 0.35922595858573914,
427
+ "learning_rate": 3.003365114974761e-07,
428
+ "loss": 0.0889,
429
+ "step": 25000
430
+ },
431
+ {
432
+ "epoch": 7.1408569028283395,
433
+ "grad_norm": 0.6694265007972717,
434
+ "learning_rate": 2.8631519910263603e-07,
435
+ "loss": 0.0904,
436
+ "step": 25500
437
+ },
438
+ {
439
+ "epoch": 7.2808737048445815,
440
+ "grad_norm": 0.41824430227279663,
441
+ "learning_rate": 2.7229388670779584e-07,
442
+ "loss": 0.0934,
443
+ "step": 26000
444
+ },
445
+ {
446
+ "epoch": 7.4208905068608235,
447
+ "grad_norm": 6.300509929656982,
448
+ "learning_rate": 2.5827257431295566e-07,
449
+ "loss": 0.0957,
450
+ "step": 26500
451
+ },
452
+ {
453
+ "epoch": 7.5609073088770655,
454
+ "grad_norm": 3.323270559310913,
455
+ "learning_rate": 2.4425126191811553e-07,
456
+ "loss": 0.0879,
457
+ "step": 27000
458
+ },
459
+ {
460
+ "epoch": 7.7009241108933075,
461
+ "grad_norm": 3.265133857727051,
462
+ "learning_rate": 2.3022994952327537e-07,
463
+ "loss": 0.091,
464
+ "step": 27500
465
+ },
466
+ {
467
+ "epoch": 7.8409409129095495,
468
+ "grad_norm": 9.813462257385254,
469
+ "learning_rate": 2.1620863712843522e-07,
470
+ "loss": 0.0895,
471
+ "step": 28000
472
+ },
473
+ {
474
+ "epoch": 7.9809577149257915,
475
+ "grad_norm": 0.7600739002227783,
476
+ "learning_rate": 2.0218732473359506e-07,
477
+ "loss": 0.0911,
478
+ "step": 28500
479
+ },
480
+ {
481
+ "epoch": 8.0,
482
+ "eval_accuracy": 0.9630519557253713,
483
+ "eval_loss": 0.1135854721069336,
484
+ "eval_model_preparation_time": 0.003,
485
+ "eval_runtime": 439.5121,
486
+ "eval_samples_per_second": 173.285,
487
+ "eval_steps_per_second": 21.663,
488
+ "step": 28568
489
+ },
490
+ {
491
+ "epoch": 8.120974516942033,
492
+ "grad_norm": 6.328824043273926,
493
+ "learning_rate": 1.881660123387549e-07,
494
+ "loss": 0.0875,
495
+ "step": 29000
496
+ },
497
+ {
498
+ "epoch": 8.260991318958276,
499
+ "grad_norm": 16.23442840576172,
500
+ "learning_rate": 1.7414469994391472e-07,
501
+ "loss": 0.089,
502
+ "step": 29500
503
+ },
504
+ {
505
+ "epoch": 8.401008120974517,
506
+ "grad_norm": 7.651858329772949,
507
+ "learning_rate": 1.601233875490746e-07,
508
+ "loss": 0.0917,
509
+ "step": 30000
510
+ },
511
+ {
512
+ "epoch": 8.54102492299076,
513
+ "grad_norm": 2.9040281772613525,
514
+ "learning_rate": 1.4610207515423443e-07,
515
+ "loss": 0.0857,
516
+ "step": 30500
517
+ },
518
+ {
519
+ "epoch": 8.681041725007,
520
+ "grad_norm": 6.958981990814209,
521
+ "learning_rate": 1.3208076275939427e-07,
522
+ "loss": 0.09,
523
+ "step": 31000
524
+ },
525
+ {
526
+ "epoch": 8.821058527023244,
527
+ "grad_norm": 4.84717321395874,
528
+ "learning_rate": 1.1805945036455412e-07,
529
+ "loss": 0.0917,
530
+ "step": 31500
531
+ },
532
+ {
533
+ "epoch": 8.961075329039485,
534
+ "grad_norm": 6.962361812591553,
535
+ "learning_rate": 1.0403813796971396e-07,
536
+ "loss": 0.0889,
537
+ "step": 32000
538
+ },
539
+ {
540
+ "epoch": 9.0,
541
+ "eval_accuracy": 0.9630256955659721,
542
+ "eval_loss": 0.11266375333070755,
543
+ "eval_model_preparation_time": 0.003,
544
+ "eval_runtime": 440.431,
545
+ "eval_samples_per_second": 172.924,
546
+ "eval_steps_per_second": 21.617,
547
+ "step": 32139
548
+ },
549
+ {
550
+ "epoch": 9.101092131055728,
551
+ "grad_norm": 13.205183982849121,
552
+ "learning_rate": 9.00168255748738e-08,
553
+ "loss": 0.09,
554
+ "step": 32500
555
+ },
556
+ {
557
+ "epoch": 9.241108933071969,
558
+ "grad_norm": 0.3649824261665344,
559
+ "learning_rate": 7.599551318003366e-08,
560
+ "loss": 0.0818,
561
+ "step": 33000
562
+ },
563
+ {
564
+ "epoch": 9.38112573508821,
565
+ "grad_norm": 1.0988820791244507,
566
+ "learning_rate": 6.197420078519349e-08,
567
+ "loss": 0.0847,
568
+ "step": 33500
569
+ },
570
+ {
571
+ "epoch": 9.521142537104453,
572
+ "grad_norm": 10.086563110351562,
573
+ "learning_rate": 4.795288839035334e-08,
574
+ "loss": 0.0909,
575
+ "step": 34000
576
+ },
577
+ {
578
+ "epoch": 9.661159339120694,
579
+ "grad_norm": 1.5432101488113403,
580
+ "learning_rate": 3.3931575995513173e-08,
581
+ "loss": 0.0873,
582
+ "step": 34500
583
+ },
584
+ {
585
+ "epoch": 9.801176141136937,
586
+ "grad_norm": 5.793679714202881,
587
+ "learning_rate": 1.9910263600673023e-08,
588
+ "loss": 0.0847,
589
+ "step": 35000
590
+ },
591
+ {
592
+ "epoch": 9.941192943153178,
593
+ "grad_norm": 1.821306824684143,
594
+ "learning_rate": 5.888951205832866e-09,
595
+ "loss": 0.0862,
596
+ "step": 35500
597
+ },
598
+ {
599
+ "epoch": 10.0,
600
+ "eval_accuracy": 0.9634327280366591,
601
+ "eval_loss": 0.11153655499219894,
602
+ "eval_model_preparation_time": 0.003,
603
+ "eval_runtime": 439.1936,
604
+ "eval_samples_per_second": 173.411,
605
+ "eval_steps_per_second": 21.678,
606
+ "step": 35710
607
+ }
608
+ ],
609
+ "logging_steps": 500,
610
+ "max_steps": 35710,
611
+ "num_input_tokens_seen": 0,
612
+ "num_train_epochs": 10,
613
+ "save_steps": 500,
614
+ "stateful_callbacks": {
615
+ "TrainerControl": {
616
+ "args": {
617
+ "should_epoch_stop": false,
618
+ "should_evaluate": false,
619
+ "should_log": false,
620
+ "should_save": true,
621
+ "should_training_stop": true
622
+ },
623
+ "attributes": {}
624
+ }
625
+ },
626
+ "total_flos": 8.852762385560602e+19,
627
+ "train_batch_size": 32,
628
+ "trial_name": null,
629
+ "trial_params": null
630
+ }
checkpoint-35710/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:803f782b12823a7c57322be2c04cedfa7e62fb7fcad0465f1e08da4bd07e8035
3
+ size 5176
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/vit-base-patch16-224-in21k",
3
+ "architectures": [
4
+ "ViTForImageClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.0,
7
+ "encoder_stride": 16,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.0,
10
+ "hidden_size": 768,
11
+ "id2label": {
12
+ "0": "Real",
13
+ "1": "Fake"
14
+ },
15
+ "image_size": 224,
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 3072,
18
+ "label2id": {
19
+ "Fake": 1,
20
+ "Real": 0
21
+ },
22
+ "layer_norm_eps": 1e-12,
23
+ "model_type": "vit",
24
+ "num_attention_heads": 12,
25
+ "num_channels": 3,
26
+ "num_hidden_layers": 12,
27
+ "patch_size": 16,
28
+ "problem_type": "single_label_classification",
29
+ "qkv_bias": true,
30
+ "torch_dtype": "float32",
31
+ "transformers_version": "4.46.0"
32
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1fd12277336e8213bb338e1897cc234131944b7594acb3751f01821dadabab9
3
+ size 343223968
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.5,
7
+ 0.5,
8
+ 0.5
9
+ ],
10
+ "image_processor_type": "ViTImageProcessor",
11
+ "image_std": [
12
+ 0.5,
13
+ 0.5,
14
+ 0.5
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 224,
20
+ "width": 224
21
+ }
22
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:803f782b12823a7c57322be2c04cedfa7e62fb7fcad0465f1e08da4bd07e8035
3
+ size 5176