nhiro3303 commited on
Commit
c3aa30b
·
1 Parent(s): 59c8ec8

Upload . with huggingface_hub

Browse files
.summary/0/events.out.tfevents.1677757888.ff5fdd36e73f ADDED
File without changes
README.md CHANGED
@@ -15,7 +15,7 @@ model-index:
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
- value: 3.62 +/- 0.56
19
  name: mean_reward
20
  verified: false
21
  ---
 
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
+ value: 4.03 +/- 0.29
19
  name: mean_reward
20
  verified: false
21
  ---
checkpoint_p0/checkpoint_000000856_3506176.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:83dd08008431c29eb9f979f192b9d1ac4662b824b7b7f7cde11cfc05d6934043
3
- size 34929220
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac7fce1699259f447856a11a89e31a0ef159b096688e88d138a6a2c799a58b45
3
+ size 34928614
checkpoint_p0/checkpoint_000000948_3883008.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b7b8d989be983293e3bf8afcd6581da7f32db06a313fbbf024d115d74315535
3
+ size 34929220
replay.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:432c1c300dd410967ba547c2ac5394f59153fe2cf5cf022f147c8bd00a8de6ca
3
- size 4879937
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d08993bc7c4a36eccde901d230a1a1eb1cf849e9bce017d1880e93ca0f0808a
3
+ size 6078630
sf_log.txt CHANGED
@@ -1139,3 +1139,292 @@ main_loop: 19.5008
1139
  [2023-03-02 11:42:40,745][09917] Avg episode rewards: #0: 3.816, true rewards: #0: 3.616
1140
  [2023-03-02 11:42:40,745][09917] Avg episode reward: 3.816, avg true_objective: 3.616
1141
  [2023-03-02 11:42:44,193][09917] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1139
  [2023-03-02 11:42:40,745][09917] Avg episode rewards: #0: 3.816, true rewards: #0: 3.616
1140
  [2023-03-02 11:42:40,745][09917] Avg episode reward: 3.816, avg true_objective: 3.616
1141
  [2023-03-02 11:42:44,193][09917] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!
1142
+ [2023-03-02 11:45:25,806][09917] The model has been pushed to https://huggingface.co/nhiro3303/rl_course_vizdoom_health_gathering_supreme
1143
+ [2023-03-02 11:51:30,436][10553] Saving configuration to /home/gpu/train_dir/default_experiment/config.json...
1144
+ [2023-03-02 11:51:30,436][10553] Rollout worker 0 uses device cpu
1145
+ [2023-03-02 11:51:30,436][10553] Rollout worker 1 uses device cpu
1146
+ [2023-03-02 11:51:30,437][10553] Rollout worker 2 uses device cpu
1147
+ [2023-03-02 11:51:30,437][10553] Rollout worker 3 uses device cpu
1148
+ [2023-03-02 11:51:30,437][10553] Rollout worker 4 uses device cpu
1149
+ [2023-03-02 11:51:30,437][10553] Rollout worker 5 uses device cpu
1150
+ [2023-03-02 11:51:30,437][10553] Rollout worker 6 uses device cpu
1151
+ [2023-03-02 11:51:30,437][10553] Rollout worker 7 uses device cpu
1152
+ [2023-03-02 11:51:30,464][10553] Using GPUs [0] for process 0 (actually maps to GPUs [0])
1153
+ [2023-03-02 11:51:30,464][10553] InferenceWorker_p0-w0: min num requests: 2
1154
+ [2023-03-02 11:51:30,481][10553] Starting all processes...
1155
+ [2023-03-02 11:51:30,481][10553] Starting process learner_proc0
1156
+ [2023-03-02 11:51:31,134][10553] Starting all processes...
1157
+ [2023-03-02 11:51:31,137][10553] Starting process inference_proc0-0
1158
+ [2023-03-02 11:51:31,137][10611] Using GPUs [0] for process 0 (actually maps to GPUs [0])
1159
+ [2023-03-02 11:51:31,137][10611] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for learning process 0
1160
+ [2023-03-02 11:51:31,137][10553] Starting process rollout_proc0
1161
+ [2023-03-02 11:51:31,137][10553] Starting process rollout_proc1
1162
+ [2023-03-02 11:51:31,138][10553] Starting process rollout_proc2
1163
+ [2023-03-02 11:51:31,138][10553] Starting process rollout_proc3
1164
+ [2023-03-02 11:51:31,138][10553] Starting process rollout_proc4
1165
+ [2023-03-02 11:51:31,140][10553] Starting process rollout_proc5
1166
+ [2023-03-02 11:51:31,142][10553] Starting process rollout_proc6
1167
+ [2023-03-02 11:51:31,142][10553] Starting process rollout_proc7
1168
+ [2023-03-02 11:51:31,173][10611] Num visible devices: 1
1169
+ [2023-03-02 11:51:31,197][10611] Starting seed is not provided
1170
+ [2023-03-02 11:51:31,197][10611] Using GPUs [0] for process 0 (actually maps to GPUs [0])
1171
+ [2023-03-02 11:51:31,197][10611] Initializing actor-critic model on device cuda:0
1172
+ [2023-03-02 11:51:31,198][10611] RunningMeanStd input shape: (3, 72, 128)
1173
+ [2023-03-02 11:51:31,199][10611] RunningMeanStd input shape: (1,)
1174
+ [2023-03-02 11:51:31,212][10611] ConvEncoder: input_channels=3
1175
+ [2023-03-02 11:51:31,450][10611] Conv encoder output size: 512
1176
+ [2023-03-02 11:51:31,451][10611] Policy head output size: 512
1177
+ [2023-03-02 11:51:31,463][10611] Created Actor Critic model with architecture:
1178
+ [2023-03-02 11:51:31,463][10611] ActorCriticSharedWeights(
1179
+ (obs_normalizer): ObservationNormalizer(
1180
+ (running_mean_std): RunningMeanStdDictInPlace(
1181
+ (running_mean_std): ModuleDict(
1182
+ (obs): RunningMeanStdInPlace()
1183
+ )
1184
+ )
1185
+ )
1186
+ (returns_normalizer): RecursiveScriptModule(original_name=RunningMeanStdInPlace)
1187
+ (encoder): VizdoomEncoder(
1188
+ (basic_encoder): ConvEncoder(
1189
+ (enc): RecursiveScriptModule(
1190
+ original_name=ConvEncoderImpl
1191
+ (conv_head): RecursiveScriptModule(
1192
+ original_name=Sequential
1193
+ (0): RecursiveScriptModule(original_name=Conv2d)
1194
+ (1): RecursiveScriptModule(original_name=ELU)
1195
+ (2): RecursiveScriptModule(original_name=Conv2d)
1196
+ (3): RecursiveScriptModule(original_name=ELU)
1197
+ (4): RecursiveScriptModule(original_name=Conv2d)
1198
+ (5): RecursiveScriptModule(original_name=ELU)
1199
+ )
1200
+ (mlp_layers): RecursiveScriptModule(
1201
+ original_name=Sequential
1202
+ (0): RecursiveScriptModule(original_name=Linear)
1203
+ (1): RecursiveScriptModule(original_name=ELU)
1204
+ )
1205
+ )
1206
+ )
1207
+ )
1208
+ (core): ModelCoreRNN(
1209
+ (core): GRU(512, 512)
1210
+ )
1211
+ (decoder): MlpDecoder(
1212
+ (mlp): Identity()
1213
+ )
1214
+ (critic_linear): Linear(in_features=512, out_features=1, bias=True)
1215
+ (action_parameterization): ActionParameterizationDefault(
1216
+ (distribution_linear): Linear(in_features=512, out_features=5, bias=True)
1217
+ )
1218
+ )
1219
+ [2023-03-02 11:51:32,207][10641] Worker 1 uses CPU cores [2, 3]
1220
+ [2023-03-02 11:51:32,226][10640] Worker 0 uses CPU cores [0, 1]
1221
+ [2023-03-02 11:51:32,280][10639] Using GPUs [0] for process 0 (actually maps to GPUs [0])
1222
+ [2023-03-02 11:51:32,280][10639] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for inference process 0
1223
+ [2023-03-02 11:51:32,287][10644] Worker 4 uses CPU cores [8, 9]
1224
+ [2023-03-02 11:51:32,287][10647] Worker 7 uses CPU cores [14, 15]
1225
+ [2023-03-02 11:51:32,289][10643] Worker 3 uses CPU cores [6, 7]
1226
+ [2023-03-02 11:51:32,289][10645] Worker 5 uses CPU cores [10, 11]
1227
+ [2023-03-02 11:51:32,295][10639] Num visible devices: 1
1228
+ [2023-03-02 11:51:32,300][10646] Worker 6 uses CPU cores [12, 13]
1229
+ [2023-03-02 11:51:32,316][10642] Worker 2 uses CPU cores [4, 5]
1230
+ [2023-03-02 11:51:33,451][10611] Using optimizer <class 'torch.optim.adam.Adam'>
1231
+ [2023-03-02 11:51:33,451][10611] Loading state from checkpoint /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000856_3506176.pth...
1232
+ [2023-03-02 11:51:33,471][10611] Loading model from checkpoint
1233
+ [2023-03-02 11:51:33,478][10611] Loaded experiment state at self.train_step=948, self.env_steps=3883008
1234
+ [2023-03-02 11:51:33,478][10611] Initialized policy 0 weights for model version 948
1235
+ [2023-03-02 11:51:33,481][10611] LearnerWorker_p0 finished initialization!
1236
+ [2023-03-02 11:51:33,482][10611] Using GPUs [0] for process 0 (actually maps to GPUs [0])
1237
+ [2023-03-02 11:51:33,588][10639] Unhandled exception CUDA error: invalid resource handle in evt loop inference_proc0-0_evt_loop
1238
+ [2023-03-02 11:51:33,897][10553] Fps is (10 sec: nan, 60 sec: nan, 300 sec: nan). Total num frames: 3883008. Throughput: 0: nan. Samples: 0. Policy #0 lag: (min: -1.0, avg: -1.0, max: -1.0)
1239
+ [2023-03-02 11:51:37,481][10553] Keyboard interrupt detected in the event loop EvtLoop [Runner_EvtLoop, process=main process 10553], exiting...
1240
+ [2023-03-02 11:51:37,481][10641] Stopping RolloutWorker_w1...
1241
+ [2023-03-02 11:51:37,481][10645] Stopping RolloutWorker_w5...
1242
+ [2023-03-02 11:51:37,481][10553] Runner profile tree view:
1243
+ main_loop: 7.0007
1244
+ [2023-03-02 11:51:37,481][10646] Stopping RolloutWorker_w6...
1245
+ [2023-03-02 11:51:37,482][10553] Collected {0: 3883008}, FPS: 0.0
1246
+ [2023-03-02 11:51:37,482][10642] Stopping RolloutWorker_w2...
1247
+ [2023-03-02 11:51:37,482][10643] Stopping RolloutWorker_w3...
1248
+ [2023-03-02 11:51:37,482][10645] Loop rollout_proc5_evt_loop terminating...
1249
+ [2023-03-02 11:51:37,482][10644] Stopping RolloutWorker_w4...
1250
+ [2023-03-02 11:51:37,482][10641] Loop rollout_proc1_evt_loop terminating...
1251
+ [2023-03-02 11:51:37,482][10647] Stopping RolloutWorker_w7...
1252
+ [2023-03-02 11:51:37,481][10611] Stopping Batcher_0...
1253
+ [2023-03-02 11:51:37,482][10643] Loop rollout_proc3_evt_loop terminating...
1254
+ [2023-03-02 11:51:37,482][10646] Loop rollout_proc6_evt_loop terminating...
1255
+ [2023-03-02 11:51:37,482][10642] Loop rollout_proc2_evt_loop terminating...
1256
+ [2023-03-02 11:51:37,482][10644] Loop rollout_proc4_evt_loop terminating...
1257
+ [2023-03-02 11:51:37,482][10647] Loop rollout_proc7_evt_loop terminating...
1258
+ [2023-03-02 11:51:37,482][10640] Stopping RolloutWorker_w0...
1259
+ [2023-03-02 11:51:37,482][10611] Loop batcher_evt_loop terminating...
1260
+ [2023-03-02 11:51:37,482][10640] Loop rollout_proc0_evt_loop terminating...
1261
+ [2023-03-02 11:51:37,483][10611] Saving /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000948_3883008.pth...
1262
+ [2023-03-02 11:51:37,527][10611] Stopping LearnerWorker_p0...
1263
+ [2023-03-02 11:51:37,527][10611] Loop learner_proc0_evt_loop terminating...
1264
+ [2023-03-02 11:51:37,566][10553] Loading existing experiment configuration from /home/gpu/train_dir/default_experiment/config.json
1265
+ [2023-03-02 11:51:37,566][10553] Overriding arg 'num_workers' with value 1 passed from command line
1266
+ [2023-03-02 11:51:37,566][10553] Adding new argument 'no_render'=True that is not in the saved config file!
1267
+ [2023-03-02 11:51:37,566][10553] Adding new argument 'save_video'=True that is not in the saved config file!
1268
+ [2023-03-02 11:51:37,566][10553] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1269
+ [2023-03-02 11:51:37,566][10553] Adding new argument 'video_name'=None that is not in the saved config file!
1270
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'max_num_frames'=1000000000.0 that is not in the saved config file!
1271
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1272
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'push_to_hub'=False that is not in the saved config file!
1273
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'hf_repository'=None that is not in the saved config file!
1274
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'policy_index'=0 that is not in the saved config file!
1275
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1276
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'train_script'=None that is not in the saved config file!
1277
+ [2023-03-02 11:51:37,567][10553] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1278
+ [2023-03-02 11:51:37,567][10553] Using frameskip 1 and render_action_repeat=4 for evaluation
1279
+ [2023-03-02 11:51:37,577][10553] Doom resolution: 160x120, resize resolution: (128, 72)
1280
+ [2023-03-02 11:51:37,577][10553] RunningMeanStd input shape: (3, 72, 128)
1281
+ [2023-03-02 11:51:37,578][10553] RunningMeanStd input shape: (1,)
1282
+ [2023-03-02 11:51:37,589][10553] ConvEncoder: input_channels=3
1283
+ [2023-03-02 11:51:37,800][10553] Conv encoder output size: 512
1284
+ [2023-03-02 11:51:37,800][10553] Policy head output size: 512
1285
+ [2023-03-02 11:51:39,438][10553] Loading state from checkpoint /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000948_3883008.pth...
1286
+ [2023-03-02 11:51:40,207][10553] Num frames 100...
1287
+ [2023-03-02 11:51:40,334][10553] Num frames 200...
1288
+ [2023-03-02 11:51:40,460][10553] Num frames 300...
1289
+ [2023-03-02 11:51:40,615][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1290
+ [2023-03-02 11:51:40,615][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1291
+ [2023-03-02 11:51:40,634][10553] Num frames 400...
1292
+ [2023-03-02 11:51:40,757][10553] Num frames 500...
1293
+ [2023-03-02 11:51:40,880][10553] Num frames 600...
1294
+ [2023-03-02 11:51:41,003][10553] Num frames 700...
1295
+ [2023-03-02 11:51:41,127][10553] Num frames 800...
1296
+ [2023-03-02 11:51:41,219][10553] Avg episode rewards: #0: 4.660, true rewards: #0: 4.160
1297
+ [2023-03-02 11:51:41,219][10553] Avg episode reward: 4.660, avg true_objective: 4.160
1298
+ [2023-03-02 11:51:41,304][10553] Num frames 900...
1299
+ [2023-03-02 11:51:41,429][10553] Num frames 1000...
1300
+ [2023-03-02 11:51:41,552][10553] Num frames 1100...
1301
+ [2023-03-02 11:51:41,677][10553] Num frames 1200...
1302
+ [2023-03-02 11:51:41,749][10553] Avg episode rewards: #0: 4.387, true rewards: #0: 4.053
1303
+ [2023-03-02 11:51:41,749][10553] Avg episode reward: 4.387, avg true_objective: 4.053
1304
+ [2023-03-02 11:51:41,854][10553] Num frames 1300...
1305
+ [2023-03-02 11:51:41,976][10553] Num frames 1400...
1306
+ [2023-03-02 11:51:42,100][10553] Num frames 1500...
1307
+ [2023-03-02 11:51:42,222][10553] Num frames 1600...
1308
+ [2023-03-02 11:51:42,355][10553] Avg episode rewards: #0: 4.660, true rewards: #0: 4.160
1309
+ [2023-03-02 11:51:42,356][10553] Avg episode reward: 4.660, avg true_objective: 4.160
1310
+ [2023-03-02 11:51:42,402][10553] Num frames 1700...
1311
+ [2023-03-02 11:51:42,527][10553] Num frames 1800...
1312
+ [2023-03-02 11:51:42,654][10553] Num frames 1900...
1313
+ [2023-03-02 11:51:42,778][10553] Num frames 2000...
1314
+ [2023-03-02 11:51:42,869][10553] Avg episode rewards: #0: 4.458, true rewards: #0: 4.058
1315
+ [2023-03-02 11:51:42,869][10553] Avg episode reward: 4.458, avg true_objective: 4.058
1316
+ [2023-03-02 11:51:42,960][10553] Num frames 2100...
1317
+ [2023-03-02 11:51:43,085][10553] Num frames 2200...
1318
+ [2023-03-02 11:51:43,213][10553] Num frames 2300...
1319
+ [2023-03-02 11:51:43,341][10553] Num frames 2400...
1320
+ [2023-03-02 11:51:43,489][10553] Avg episode rewards: #0: 4.628, true rewards: #0: 4.128
1321
+ [2023-03-02 11:51:43,490][10553] Avg episode reward: 4.628, avg true_objective: 4.128
1322
+ [2023-03-02 11:51:43,517][10553] Num frames 2500...
1323
+ [2023-03-02 11:51:43,642][10553] Num frames 2600...
1324
+ [2023-03-02 11:51:43,766][10553] Num frames 2700...
1325
+ [2023-03-02 11:51:43,890][10553] Num frames 2800...
1326
+ [2023-03-02 11:51:44,014][10553] Num frames 2900...
1327
+ [2023-03-02 11:51:44,193][10553] Avg episode rewards: #0: 5.127, true rewards: #0: 4.270
1328
+ [2023-03-02 11:51:44,193][10553] Avg episode reward: 5.127, avg true_objective: 4.270
1329
+ [2023-03-02 11:51:44,207][10553] Num frames 3000...
1330
+ [2023-03-02 11:51:44,336][10553] Num frames 3100...
1331
+ [2023-03-02 11:51:44,460][10553] Num frames 3200...
1332
+ [2023-03-02 11:51:44,582][10553] Num frames 3300...
1333
+ [2023-03-02 11:51:44,726][10553] Avg episode rewards: #0: 4.966, true rewards: #0: 4.216
1334
+ [2023-03-02 11:51:44,726][10553] Avg episode reward: 4.966, avg true_objective: 4.216
1335
+ [2023-03-02 11:51:44,764][10553] Num frames 3400...
1336
+ [2023-03-02 11:51:44,910][10553] Num frames 3500...
1337
+ [2023-03-02 11:51:45,053][10553] Num frames 3600...
1338
+ [2023-03-02 11:51:45,198][10553] Num frames 3700...
1339
+ [2023-03-02 11:51:45,350][10553] Num frames 3800...
1340
+ [2023-03-02 11:51:45,438][10553] Avg episode rewards: #0: 5.023, true rewards: #0: 4.246
1341
+ [2023-03-02 11:51:45,438][10553] Avg episode reward: 5.023, avg true_objective: 4.246
1342
+ [2023-03-02 11:51:45,553][10553] Num frames 3900...
1343
+ [2023-03-02 11:51:45,697][10553] Num frames 4000...
1344
+ [2023-03-02 11:51:45,846][10553] Num frames 4100...
1345
+ [2023-03-02 11:51:46,005][10553] Num frames 4200...
1346
+ [2023-03-02 11:51:46,068][10553] Avg episode rewards: #0: 4.905, true rewards: #0: 4.205
1347
+ [2023-03-02 11:51:46,068][10553] Avg episode reward: 4.905, avg true_objective: 4.205
1348
+ [2023-03-02 11:51:50,116][10553] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!
1349
+ [2023-03-02 11:52:10,809][10553] Loading existing experiment configuration from /home/gpu/train_dir/default_experiment/config.json
1350
+ [2023-03-02 11:52:10,809][10553] Overriding arg 'num_workers' with value 1 passed from command line
1351
+ [2023-03-02 11:52:10,809][10553] Adding new argument 'no_render'=True that is not in the saved config file!
1352
+ [2023-03-02 11:52:10,809][10553] Adding new argument 'save_video'=True that is not in the saved config file!
1353
+ [2023-03-02 11:52:10,809][10553] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1354
+ [2023-03-02 11:52:10,809][10553] Adding new argument 'video_name'=None that is not in the saved config file!
1355
+ [2023-03-02 11:52:10,809][10553] Adding new argument 'max_num_frames'=100000 that is not in the saved config file!
1356
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1357
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'push_to_hub'=True that is not in the saved config file!
1358
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'hf_repository'='nhiro3303/rl_course_vizdoom_health_gathering_supreme' that is not in the saved config file!
1359
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'policy_index'=0 that is not in the saved config file!
1360
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1361
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'train_script'=None that is not in the saved config file!
1362
+ [2023-03-02 11:52:10,810][10553] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1363
+ [2023-03-02 11:52:10,810][10553] Using frameskip 1 and render_action_repeat=4 for evaluation
1364
+ [2023-03-02 11:52:10,813][10553] RunningMeanStd input shape: (3, 72, 128)
1365
+ [2023-03-02 11:52:10,814][10553] RunningMeanStd input shape: (1,)
1366
+ [2023-03-02 11:52:10,821][10553] ConvEncoder: input_channels=3
1367
+ [2023-03-02 11:52:10,844][10553] Conv encoder output size: 512
1368
+ [2023-03-02 11:52:10,845][10553] Policy head output size: 512
1369
+ [2023-03-02 11:52:10,864][10553] Loading state from checkpoint /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000948_3883008.pth...
1370
+ [2023-03-02 11:52:11,229][10553] Num frames 100...
1371
+ [2023-03-02 11:52:11,422][10553] Num frames 200...
1372
+ [2023-03-02 11:52:11,622][10553] Num frames 300...
1373
+ [2023-03-02 11:52:11,837][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1374
+ [2023-03-02 11:52:11,837][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1375
+ [2023-03-02 11:52:11,873][10553] Num frames 400...
1376
+ [2023-03-02 11:52:12,065][10553] Num frames 500...
1377
+ [2023-03-02 11:52:12,256][10553] Num frames 600...
1378
+ [2023-03-02 11:52:12,452][10553] Num frames 700...
1379
+ [2023-03-02 11:52:12,641][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1380
+ [2023-03-02 11:52:12,641][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1381
+ [2023-03-02 11:52:12,710][10553] Num frames 800...
1382
+ [2023-03-02 11:52:12,902][10553] Num frames 900...
1383
+ [2023-03-02 11:52:13,095][10553] Num frames 1000...
1384
+ [2023-03-02 11:52:13,289][10553] Num frames 1100...
1385
+ [2023-03-02 11:52:13,456][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1386
+ [2023-03-02 11:52:13,456][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1387
+ [2023-03-02 11:52:13,557][10553] Num frames 1200...
1388
+ [2023-03-02 11:52:13,745][10553] Num frames 1300...
1389
+ [2023-03-02 11:52:13,935][10553] Num frames 1400...
1390
+ [2023-03-02 11:52:14,136][10553] Num frames 1500...
1391
+ [2023-03-02 11:52:14,270][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1392
+ [2023-03-02 11:52:14,271][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1393
+ [2023-03-02 11:52:14,406][10553] Num frames 1600...
1394
+ [2023-03-02 11:52:14,609][10553] Num frames 1700...
1395
+ [2023-03-02 11:52:14,802][10553] Num frames 1800...
1396
+ [2023-03-02 11:52:14,994][10553] Num frames 1900...
1397
+ [2023-03-02 11:52:15,091][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1398
+ [2023-03-02 11:52:15,091][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1399
+ [2023-03-02 11:52:15,246][10553] Num frames 2000...
1400
+ [2023-03-02 11:52:15,446][10553] Num frames 2100...
1401
+ [2023-03-02 11:52:15,640][10553] Num frames 2200...
1402
+ [2023-03-02 11:52:15,835][10553] Num frames 2300...
1403
+ [2023-03-02 11:52:15,895][10553] Avg episode rewards: #0: 3.840, true rewards: #0: 3.840
1404
+ [2023-03-02 11:52:15,896][10553] Avg episode reward: 3.840, avg true_objective: 3.840
1405
+ [2023-03-02 11:52:16,085][10553] Num frames 2400...
1406
+ [2023-03-02 11:52:16,278][10553] Num frames 2500...
1407
+ [2023-03-02 11:52:16,466][10553] Num frames 2600...
1408
+ [2023-03-02 11:52:16,654][10553] Num frames 2700...
1409
+ [2023-03-02 11:52:16,812][10553] Avg episode rewards: #0: 4.074, true rewards: #0: 3.931
1410
+ [2023-03-02 11:52:16,813][10553] Avg episode reward: 4.074, avg true_objective: 3.931
1411
+ [2023-03-02 11:52:16,907][10553] Num frames 2800...
1412
+ [2023-03-02 11:52:17,096][10553] Num frames 2900...
1413
+ [2023-03-02 11:52:17,285][10553] Num frames 3000...
1414
+ [2023-03-02 11:52:17,478][10553] Num frames 3100...
1415
+ [2023-03-02 11:52:17,609][10553] Avg episode rewards: #0: 4.045, true rewards: #0: 3.920
1416
+ [2023-03-02 11:52:17,609][10553] Avg episode reward: 4.045, avg true_objective: 3.920
1417
+ [2023-03-02 11:52:17,737][10553] Num frames 3200...
1418
+ [2023-03-02 11:52:17,927][10553] Num frames 3300...
1419
+ [2023-03-02 11:52:18,122][10553] Num frames 3400...
1420
+ [2023-03-02 11:52:18,317][10553] Num frames 3500...
1421
+ [2023-03-02 11:52:18,538][10553] Avg episode rewards: #0: 4.204, true rewards: #0: 3.982
1422
+ [2023-03-02 11:52:18,538][10553] Avg episode reward: 4.204, avg true_objective: 3.982
1423
+ [2023-03-02 11:52:18,577][10553] Num frames 3600...
1424
+ [2023-03-02 11:52:18,769][10553] Num frames 3700...
1425
+ [2023-03-02 11:52:18,965][10553] Num frames 3800...
1426
+ [2023-03-02 11:52:19,169][10553] Num frames 3900...
1427
+ [2023-03-02 11:52:19,370][10553] Num frames 4000...
1428
+ [2023-03-02 11:52:19,493][10553] Avg episode rewards: #0: 4.332, true rewards: #0: 4.032
1429
+ [2023-03-02 11:52:19,494][10553] Avg episode reward: 4.332, avg true_objective: 4.032
1430
+ [2023-03-02 11:52:23,407][10553] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!