RajkNakka commited on
Commit
94ddc31
1 Parent(s): f3c6b3f

Upload folder using huggingface_hub

Browse files
.summary/0/events.out.tfevents.1688872419.snowflake ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:745332bab2e2aeabae1e5c353576ade83586f1d5a76bb3136a0fe6d20aaf5a27
3
+ size 15236
.summary/0/events.out.tfevents.1688873967.snowflake ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90cbaf7dc8e5fd40d00d33dfae5ed26b91d56fdddb7b11908f9de42cb44e425e
3
+ size 15236
README.md CHANGED
@@ -15,7 +15,7 @@ model-index:
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
- value: 4.06 +/- 0.29
19
  name: mean_reward
20
  verified: false
21
  ---
 
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
+ value: 3.88 +/- 0.22
19
  name: mean_reward
20
  verified: false
21
  ---
config.json CHANGED
@@ -4,7 +4,7 @@
4
  "env": "doom_health_gathering_supreme",
5
  "experiment": "default_experiment",
6
  "train_dir": "/home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir",
7
- "restart_behavior": "overwrite",
8
  "device": "gpu",
9
  "seed": 42,
10
  "num_policies": 1,
@@ -130,8 +130,8 @@
130
  "cli_args": {},
131
  "git_hash": "0401714b01ee832562a0930e3744117f1ba51e10",
132
  "git_repo_name": "https://github.com/tenkara/HF-DeepRL.git",
 
133
  "lr_adaptive_min": 1e-06,
134
  "lr_adaptive_max": 0.01,
135
- "env_gpu_observations": true,
136
- "command_line": "--env=doom_health_gathering_supreme --train_for_env_steps=2000000"
137
  }
 
4
  "env": "doom_health_gathering_supreme",
5
  "experiment": "default_experiment",
6
  "train_dir": "/home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir",
7
+ "restart_behavior": "resume",
8
  "device": "gpu",
9
  "seed": 42,
10
  "num_policies": 1,
 
130
  "cli_args": {},
131
  "git_hash": "0401714b01ee832562a0930e3744117f1ba51e10",
132
  "git_repo_name": "https://github.com/tenkara/HF-DeepRL.git",
133
+ "command_line": "--env=doom_health_gathering_supreme --train_for_env_steps=2000000",
134
  "lr_adaptive_min": 1e-06,
135
  "lr_adaptive_max": 0.01,
136
+ "env_gpu_observations": true
 
137
  }
git.diff CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bffe781f41b26f0bcb140521428a958667eb1efa3e48466573f44ef778c97965
3
- size 212386620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a61a193f0cc481d66bfc2627e32f7543022b6969516b6c1e19e74af543e7c41
3
+ size 212435236
sf_log.txt CHANGED
@@ -138,3 +138,285 @@ Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
138
  [2023-07-08 20:59:39,043][17004] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
139
  [2023-07-08 20:59:39,075][17004] Stopping LearnerWorker_p0...
140
  [2023-07-08 20:59:39,075][17004] Loop learner_proc0_evt_loop terminating...
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
138
  [2023-07-08 20:59:39,043][17004] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
139
  [2023-07-08 20:59:39,075][17004] Stopping LearnerWorker_p0...
140
  [2023-07-08 20:59:39,075][17004] Loop learner_proc0_evt_loop terminating...
141
+ [2023-07-08 21:13:45,642][17306] Using GPUs [0] for process 0 (actually maps to GPUs [0])
142
+ [2023-07-08 21:13:45,642][17306] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for learning process 0
143
+ [2023-07-08 21:13:45,688][17306] Num visible devices: 1
144
+ [2023-07-08 21:13:45,807][17306] Setting fixed seed 42
145
+ [2023-07-08 21:13:45,808][17306] Using GPUs [0] for process 0 (actually maps to GPUs [0])
146
+ [2023-07-08 21:13:45,808][17306] Initializing actor-critic model on device cuda:0
147
+ [2023-07-08 21:13:45,808][17306] RunningMeanStd input shape: (3, 72, 128)
148
+ [2023-07-08 21:13:45,809][17306] RunningMeanStd input shape: (1,)
149
+ [2023-07-08 21:13:45,816][17306] ConvEncoder: input_channels=3
150
+ [2023-07-08 21:13:45,919][17326] Using GPUs [0] for process 0 (actually maps to GPUs [0])
151
+ [2023-07-08 21:13:45,920][17326] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for inference process 0
152
+ [2023-07-08 21:13:45,976][17326] Num visible devices: 1
153
+ [2023-07-08 21:13:46,140][17327] Worker 0 uses CPU cores [0]
154
+ [2023-07-08 21:13:46,130][17306] Conv encoder output size: 512
155
+ [2023-07-08 21:13:46,172][17306] Policy head output size: 512
156
+ [2023-07-08 21:13:46,223][17306] Created Actor Critic model with architecture:
157
+ [2023-07-08 21:13:46,242][17306] ActorCriticSharedWeights(
158
+ (obs_normalizer): ObservationNormalizer(
159
+ (running_mean_std): RunningMeanStdDictInPlace(
160
+ (running_mean_std): ModuleDict(
161
+ (obs): RunningMeanStdInPlace()
162
+ )
163
+ )
164
+ )
165
+ (returns_normalizer): RecursiveScriptModule(original_name=RunningMeanStdInPlace)
166
+ (encoder): VizdoomEncoder(
167
+ (basic_encoder): ConvEncoder(
168
+ (enc): RecursiveScriptModule(
169
+ original_name=ConvEncoderImpl
170
+ (conv_head): RecursiveScriptModule(
171
+ original_name=Sequential
172
+ (0): RecursiveScriptModule(original_name=Conv2d)
173
+ (1): RecursiveScriptModule(original_name=ReLU)
174
+ (2): RecursiveScriptModule(original_name=Conv2d)
175
+ (3): RecursiveScriptModule(original_name=ReLU)
176
+ (4): RecursiveScriptModule(original_name=Conv2d)
177
+ (5): RecursiveScriptModule(original_name=ReLU)
178
+ )
179
+ (mlp_layers): RecursiveScriptModule(
180
+ original_name=Sequential
181
+ (0): RecursiveScriptModule(original_name=Linear)
182
+ (1): RecursiveScriptModule(original_name=ReLU)
183
+ )
184
+ )
185
+ )
186
+ )
187
+ (core): ModelCoreRNN(
188
+ (core): LSTM(512, 512)
189
+ )
190
+ (decoder): MlpDecoder(
191
+ (mlp): Identity()
192
+ )
193
+ (critic_linear): Linear(in_features=512, out_features=1, bias=True)
194
+ (action_parameterization): ActionParameterizationDefault(
195
+ (distribution_linear): Linear(in_features=512, out_features=5, bias=True)
196
+ )
197
+ )
198
+ [2023-07-08 21:13:46,351][17329] Worker 2 uses CPU cores [2]
199
+ [2023-07-08 21:13:46,421][17330] Worker 3 uses CPU cores [3]
200
+ [2023-07-08 21:13:46,561][17331] Worker 4 uses CPU cores [0]
201
+ [2023-07-08 21:13:46,608][17334] Worker 5 uses CPU cores [1]
202
+ [2023-07-08 21:13:46,689][17338] Worker 11 uses CPU cores [3]
203
+ [2023-07-08 21:13:46,697][17336] Worker 9 uses CPU cores [1]
204
+ [2023-07-08 21:13:46,698][17339] Worker 12 uses CPU cores [0]
205
+ [2023-07-08 21:13:46,704][17328] Worker 1 uses CPU cores [1]
206
+ [2023-07-08 21:13:46,711][17345] Worker 18 uses CPU cores [2]
207
+ [2023-07-08 21:13:46,732][17346] Worker 19 uses CPU cores [3]
208
+ [2023-07-08 21:13:46,743][17341] Worker 15 uses CPU cores [3]
209
+ [2023-07-08 21:13:46,746][17340] Worker 13 uses CPU cores [1]
210
+ [2023-07-08 21:13:46,746][17344] Worker 17 uses CPU cores [1]
211
+ [2023-07-08 21:13:46,771][17337] Worker 10 uses CPU cores [2]
212
+ [2023-07-08 21:13:46,789][17333] Worker 7 uses CPU cores [3]
213
+ [2023-07-08 21:13:46,791][17332] Worker 6 uses CPU cores [2]
214
+ [2023-07-08 21:13:46,793][17343] Worker 16 uses CPU cores [0]
215
+ [2023-07-08 21:13:46,801][17335] Worker 8 uses CPU cores [0]
216
+ [2023-07-08 21:13:46,841][17342] Worker 14 uses CPU cores [2]
217
+ [2023-07-08 21:13:47,146][17306] Using optimizer <class 'torch.optim.adam.Adam'>
218
+ [2023-07-08 21:13:47,147][17306] Loading state from checkpoint /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
219
+ [2023-07-08 21:13:47,155][17306] Loading model from checkpoint
220
+ [2023-07-08 21:13:47,156][17306] Loaded experiment state at self.train_step=0, self.env_steps=0
221
+ [2023-07-08 21:13:47,156][17306] Initialized policy 0 weights for model version 0
222
+ [2023-07-08 21:13:47,159][17306] Using GPUs [0] for process 0 (actually maps to GPUs [0])
223
+ [2023-07-08 21:13:47,161][17306] LearnerWorker_p0 finished initialization!
224
+ [2023-07-08 21:13:47,303][17326] Unhandled exception CUDA error: OS call failed or operation not supported on this OS
225
+ CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
226
+ For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
227
+ Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
228
+ in evt loop inference_proc0-0_evt_loop
229
+ [2023-07-08 21:15:39,049][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
230
+ [2023-07-08 21:17:39,050][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
231
+ [2023-07-08 21:19:39,050][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
232
+ [2023-07-08 21:21:39,050][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
233
+ [2023-07-08 21:23:39,050][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
234
+ [2023-07-08 21:25:39,050][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
235
+ [2023-07-08 21:27:39,050][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
236
+ [2023-07-08 21:28:39,052][17337] Stopping RolloutWorker_w10...
237
+ [2023-07-08 21:28:39,052][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
238
+ [2023-07-08 21:28:39,052][17337] Loop rollout_proc10_evt_loop terminating...
239
+ [2023-07-08 21:28:39,052][17335] Stopping RolloutWorker_w8...
240
+ [2023-07-08 21:28:39,053][17339] Stopping RolloutWorker_w12...
241
+ [2023-07-08 21:28:39,054][17333] Stopping RolloutWorker_w7...
242
+ [2023-07-08 21:28:39,053][17343] Stopping RolloutWorker_w16...
243
+ [2023-07-08 21:28:39,055][17341] Stopping RolloutWorker_w15...
244
+ [2023-07-08 21:28:39,055][17343] Loop rollout_proc16_evt_loop terminating...
245
+ [2023-07-08 21:28:39,054][17346] Stopping RolloutWorker_w19...
246
+ [2023-07-08 21:28:39,055][17338] Stopping RolloutWorker_w11...
247
+ [2023-07-08 21:28:39,055][17330] Stopping RolloutWorker_w3...
248
+ [2023-07-08 21:28:39,056][17341] Loop rollout_proc15_evt_loop terminating...
249
+ [2023-07-08 21:28:39,053][17331] Stopping RolloutWorker_w4...
250
+ [2023-07-08 21:28:39,056][17333] Loop rollout_proc7_evt_loop terminating...
251
+ [2023-07-08 21:28:39,057][17331] Loop rollout_proc4_evt_loop terminating...
252
+ [2023-07-08 21:28:39,056][17338] Loop rollout_proc11_evt_loop terminating...
253
+ [2023-07-08 21:28:39,061][17345] Stopping RolloutWorker_w18...
254
+ [2023-07-08 21:28:39,061][17306] Stopping Batcher_0...
255
+ [2023-07-08 21:28:39,061][17345] Loop rollout_proc18_evt_loop terminating...
256
+ [2023-07-08 21:28:39,061][17306] Loop batcher_evt_loop terminating...
257
+ [2023-07-08 21:28:39,061][17335] Loop rollout_proc8_evt_loop terminating...
258
+ [2023-07-08 21:28:39,057][17346] Loop rollout_proc19_evt_loop terminating...
259
+ [2023-07-08 21:28:39,062][17340] Stopping RolloutWorker_w13...
260
+ [2023-07-08 21:28:39,057][17330] Loop rollout_proc3_evt_loop terminating...
261
+ [2023-07-08 21:28:39,062][17340] Loop rollout_proc13_evt_loop terminating...
262
+ [2023-07-08 21:28:39,071][17329] Stopping RolloutWorker_w2...
263
+ [2023-07-08 21:28:39,071][17329] Loop rollout_proc2_evt_loop terminating...
264
+ [2023-07-08 21:28:39,052][17327] Stopping RolloutWorker_w0...
265
+ [2023-07-08 21:28:39,072][17344] Stopping RolloutWorker_w17...
266
+ [2023-07-08 21:28:39,072][17327] Loop rollout_proc0_evt_loop terminating...
267
+ [2023-07-08 21:28:39,072][17344] Loop rollout_proc17_evt_loop terminating...
268
+ [2023-07-08 21:28:39,081][17332] Stopping RolloutWorker_w6...
269
+ [2023-07-08 21:28:39,081][17332] Loop rollout_proc6_evt_loop terminating...
270
+ [2023-07-08 21:28:39,071][17339] Loop rollout_proc12_evt_loop terminating...
271
+ [2023-07-08 21:28:39,082][17328] Stopping RolloutWorker_w1...
272
+ [2023-07-08 21:28:39,082][17328] Loop rollout_proc1_evt_loop terminating...
273
+ [2023-07-08 21:28:39,091][17342] Stopping RolloutWorker_w14...
274
+ [2023-07-08 21:28:39,091][17342] Loop rollout_proc14_evt_loop terminating...
275
+ [2023-07-08 21:28:39,092][17334] Stopping RolloutWorker_w5...
276
+ [2023-07-08 21:28:39,092][17334] Loop rollout_proc5_evt_loop terminating...
277
+ [2023-07-08 21:28:39,094][17306] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
278
+ [2023-07-08 21:28:39,102][17336] Stopping RolloutWorker_w9...
279
+ [2023-07-08 21:28:39,102][17336] Loop rollout_proc9_evt_loop terminating...
280
+ [2023-07-08 21:28:39,125][17306] Stopping LearnerWorker_p0...
281
+ [2023-07-08 21:28:39,125][17306] Loop learner_proc0_evt_loop terminating...
282
+ [2023-07-08 21:39:33,872][17857] Using GPUs [0] for process 0 (actually maps to GPUs [0])
283
+ [2023-07-08 21:39:33,882][17857] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for learning process 0
284
+ [2023-07-08 21:39:33,928][17857] Num visible devices: 1
285
+ [2023-07-08 21:39:34,048][17857] Setting fixed seed 42
286
+ [2023-07-08 21:39:34,049][17857] Using GPUs [0] for process 0 (actually maps to GPUs [0])
287
+ [2023-07-08 21:39:34,049][17857] Initializing actor-critic model on device cuda:0
288
+ [2023-07-08 21:39:34,049][17857] RunningMeanStd input shape: (3, 72, 128)
289
+ [2023-07-08 21:39:34,049][17857] RunningMeanStd input shape: (1,)
290
+ [2023-07-08 21:39:34,056][17857] ConvEncoder: input_channels=3
291
+ [2023-07-08 21:39:34,442][17857] Conv encoder output size: 512
292
+ [2023-07-08 21:39:34,443][17857] Policy head output size: 512
293
+ [2023-07-08 21:39:34,452][17857] Created Actor Critic model with architecture:
294
+ [2023-07-08 21:39:34,492][17857] ActorCriticSharedWeights(
295
+ (obs_normalizer): ObservationNormalizer(
296
+ (running_mean_std): RunningMeanStdDictInPlace(
297
+ (running_mean_std): ModuleDict(
298
+ (obs): RunningMeanStdInPlace()
299
+ )
300
+ )
301
+ )
302
+ (returns_normalizer): RecursiveScriptModule(original_name=RunningMeanStdInPlace)
303
+ (encoder): VizdoomEncoder(
304
+ (basic_encoder): ConvEncoder(
305
+ (enc): RecursiveScriptModule(
306
+ original_name=ConvEncoderImpl
307
+ (conv_head): RecursiveScriptModule(
308
+ original_name=Sequential
309
+ (0): RecursiveScriptModule(original_name=Conv2d)
310
+ (1): RecursiveScriptModule(original_name=ReLU)
311
+ (2): RecursiveScriptModule(original_name=Conv2d)
312
+ (3): RecursiveScriptModule(original_name=ReLU)
313
+ (4): RecursiveScriptModule(original_name=Conv2d)
314
+ (5): RecursiveScriptModule(original_name=ReLU)
315
+ )
316
+ (mlp_layers): RecursiveScriptModule(
317
+ original_name=Sequential
318
+ (0): RecursiveScriptModule(original_name=Linear)
319
+ (1): RecursiveScriptModule(original_name=ReLU)
320
+ )
321
+ )
322
+ )
323
+ )
324
+ (core): ModelCoreRNN(
325
+ (core): LSTM(512, 512)
326
+ )
327
+ (decoder): MlpDecoder(
328
+ (mlp): Identity()
329
+ )
330
+ (critic_linear): Linear(in_features=512, out_features=1, bias=True)
331
+ (action_parameterization): ActionParameterizationDefault(
332
+ (distribution_linear): Linear(in_features=512, out_features=5, bias=True)
333
+ )
334
+ )
335
+ [2023-07-08 21:39:34,999][17857] Using optimizer <class 'torch.optim.adam.Adam'>
336
+ [2023-07-08 21:39:35,000][17857] Loading state from checkpoint /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
337
+ [2023-07-08 21:39:35,072][17857] Loading model from checkpoint
338
+ [2023-07-08 21:39:35,073][17857] Loaded experiment state at self.train_step=0, self.env_steps=0
339
+ [2023-07-08 21:39:35,074][17857] Initialized policy 0 weights for model version 0
340
+ [2023-07-08 21:39:35,084][17857] Using GPUs [0] for process 0 (actually maps to GPUs [0])
341
+ [2023-07-08 21:39:35,092][17857] LearnerWorker_p0 finished initialization!
342
+ [2023-07-08 21:39:35,100][17884] Worker 6 uses CPU cores [2]
343
+ [2023-07-08 21:39:35,221][17877] Using GPUs [0] for process 0 (actually maps to GPUs [0])
344
+ [2023-07-08 21:39:35,222][17877] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for inference process 0
345
+ [2023-07-08 21:39:35,223][17878] Worker 0 uses CPU cores [0]
346
+ [2023-07-08 21:39:35,231][17879] Worker 1 uses CPU cores [1]
347
+ [2023-07-08 21:39:35,286][17877] Num visible devices: 1
348
+ [2023-07-08 21:39:35,371][17880] Worker 2 uses CPU cores [2]
349
+ [2023-07-08 21:39:35,398][17886] Worker 8 uses CPU cores [0]
350
+ [2023-07-08 21:39:35,408][17888] Worker 10 uses CPU cores [2]
351
+ [2023-07-08 21:39:35,492][17885] Worker 7 uses CPU cores [3]
352
+ [2023-07-08 21:39:35,492][17891] Worker 12 uses CPU cores [0]
353
+ [2023-07-08 21:39:35,511][17887] Worker 9 uses CPU cores [1]
354
+ [2023-07-08 21:39:35,532][17881] Worker 4 uses CPU cores [0]
355
+ [2023-07-08 21:39:35,541][17882] Worker 3 uses CPU cores [3]
356
+ [2023-07-08 21:39:35,576][17893] Worker 15 uses CPU cores [3]
357
+ [2023-07-08 21:39:35,601][17890] Worker 13 uses CPU cores [1]
358
+ [2023-07-08 21:39:35,613][17892] Worker 14 uses CPU cores [2]
359
+ [2023-07-08 21:39:35,621][17894] Worker 16 uses CPU cores [0]
360
+ [2023-07-08 21:39:35,624][17895] Worker 17 uses CPU cores [1]
361
+ [2023-07-08 21:39:35,672][17896] Worker 19 uses CPU cores [3]
362
+ [2023-07-08 21:39:35,679][17883] Worker 5 uses CPU cores [1]
363
+ [2023-07-08 21:39:35,686][17897] Worker 18 uses CPU cores [2]
364
+ [2023-07-08 21:39:35,725][17889] Worker 11 uses CPU cores [3]
365
+ [2023-07-08 21:39:35,840][17877] Unhandled exception CUDA error: OS call failed or operation not supported on this OS
366
+ CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
367
+ For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
368
+ Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
369
+ in evt loop inference_proc0-0_evt_loop
370
+ [2023-07-08 21:41:27,794][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
371
+ [2023-07-08 21:43:27,795][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
372
+ [2023-07-08 21:45:27,793][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
373
+ [2023-07-08 21:47:27,793][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
374
+ [2023-07-08 21:49:27,793][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
375
+ [2023-07-08 21:51:27,793][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
376
+ [2023-07-08 21:53:27,795][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
377
+ [2023-07-08 21:54:27,795][17894] Stopping RolloutWorker_w16...
378
+ [2023-07-08 21:54:27,795][17890] Stopping RolloutWorker_w13...
379
+ [2023-07-08 21:54:27,795][17894] Loop rollout_proc16_evt_loop terminating...
380
+ [2023-07-08 21:54:27,795][17890] Loop rollout_proc13_evt_loop terminating...
381
+ [2023-07-08 21:54:27,796][17888] Stopping RolloutWorker_w10...
382
+ [2023-07-08 21:54:27,796][17888] Loop rollout_proc10_evt_loop terminating...
383
+ [2023-07-08 21:54:27,802][17879] Stopping RolloutWorker_w1...
384
+ [2023-07-08 21:54:27,802][17892] Stopping RolloutWorker_w14...
385
+ [2023-07-08 21:54:27,802][17891] Stopping RolloutWorker_w12...
386
+ [2023-07-08 21:54:27,802][17896] Stopping RolloutWorker_w19...
387
+ [2023-07-08 21:54:27,802][17879] Loop rollout_proc1_evt_loop terminating...
388
+ [2023-07-08 21:54:27,802][17892] Loop rollout_proc14_evt_loop terminating...
389
+ [2023-07-08 21:54:27,802][17891] Loop rollout_proc12_evt_loop terminating...
390
+ [2023-07-08 21:54:27,802][17896] Loop rollout_proc19_evt_loop terminating...
391
+ [2023-07-08 21:54:27,808][17897] Stopping RolloutWorker_w18...
392
+ [2023-07-08 21:54:27,808][17897] Loop rollout_proc18_evt_loop terminating...
393
+ [2023-07-08 21:54:27,812][17885] Stopping RolloutWorker_w7...
394
+ [2023-07-08 21:54:27,812][17881] Stopping RolloutWorker_w4...
395
+ [2023-07-08 21:54:27,812][17880] Stopping RolloutWorker_w2...
396
+ [2023-07-08 21:54:27,812][17885] Loop rollout_proc7_evt_loop terminating...
397
+ [2023-07-08 21:54:27,812][17895] Stopping RolloutWorker_w17...
398
+ [2023-07-08 21:54:27,812][17881] Loop rollout_proc4_evt_loop terminating...
399
+ [2023-07-08 21:54:27,812][17880] Loop rollout_proc2_evt_loop terminating...
400
+ [2023-07-08 21:54:27,812][17895] Loop rollout_proc17_evt_loop terminating...
401
+ [2023-07-08 21:54:27,818][17884] Stopping RolloutWorker_w6...
402
+ [2023-07-08 21:54:27,819][17884] Loop rollout_proc6_evt_loop terminating...
403
+ [2023-07-08 21:54:27,822][17893] Stopping RolloutWorker_w15...
404
+ [2023-07-08 21:54:27,822][17883] Stopping RolloutWorker_w5...
405
+ [2023-07-08 21:54:27,822][17878] Stopping RolloutWorker_w0...
406
+ [2023-07-08 21:54:27,822][17893] Loop rollout_proc15_evt_loop terminating...
407
+ [2023-07-08 21:54:27,822][17883] Loop rollout_proc5_evt_loop terminating...
408
+ [2023-07-08 21:54:27,822][17878] Loop rollout_proc0_evt_loop terminating...
409
+ [2023-07-08 21:54:27,832][17882] Stopping RolloutWorker_w3...
410
+ [2023-07-08 21:54:27,832][17882] Loop rollout_proc3_evt_loop terminating...
411
+ [2023-07-08 21:54:27,832][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
412
+ [2023-07-08 21:54:27,833][17886] Stopping RolloutWorker_w8...
413
+ [2023-07-08 21:54:27,833][17886] Loop rollout_proc8_evt_loop terminating...
414
+ [2023-07-08 21:54:27,833][17887] Stopping RolloutWorker_w9...
415
+ [2023-07-08 21:54:27,833][17887] Loop rollout_proc9_evt_loop terminating...
416
+ [2023-07-08 21:54:27,839][17857] Stopping Batcher_0...
417
+ [2023-07-08 21:54:27,839][17857] Loop batcher_evt_loop terminating...
418
+ [2023-07-08 21:54:27,842][17889] Stopping RolloutWorker_w11...
419
+ [2023-07-08 21:54:27,842][17889] Loop rollout_proc11_evt_loop terminating...
420
+ [2023-07-08 21:54:27,855][17857] Saving /home/raj/repos/HF-DeepRL/8-Proximal-Policy-Optimization/train_dir/default_experiment/checkpoint_p0/checkpoint_000000000_0.pth...
421
+ [2023-07-08 21:54:27,905][17857] Stopping LearnerWorker_p0...
422
+ [2023-07-08 21:54:27,905][17857] Loop learner_proc0_evt_loop terminating...