nhiro3303 commited on
Commit
59c8ec8
·
1 Parent(s): 5ec7676

Upload . with huggingface_hub

Browse files
.summary/0/events.out.tfevents.1677757307.ff5fdd36e73f ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd97fbae60f5186db877d0d1680b08a0c9635120ef51163b4ad617394622de11
3
+ size 410
README.md CHANGED
@@ -15,7 +15,7 @@ model-index:
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
- value: 4.06 +/- 0.29
19
  name: mean_reward
20
  verified: false
21
  ---
 
15
  type: doom_health_gathering_supreme
16
  metrics:
17
  - type: mean_reward
18
+ value: 3.62 +/- 0.56
19
  name: mean_reward
20
  verified: false
21
  ---
checkpoint_p0/checkpoint_000000856_3506176.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83dd08008431c29eb9f979f192b9d1ac4662b824b7b7f7cde11cfc05d6934043
3
+ size 34929220
replay.mp4 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:481723c87624e186b57cc803e7ddac7e31e2d73bb360d1e9469344597411336c
3
- size 5837073
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:432c1c300dd410967ba547c2ac5394f59153fe2cf5cf022f147c8bd00a8de6ca
3
+ size 4879937
sf_log.txt CHANGED
@@ -853,3 +853,289 @@ main_loop: 967.7497
853
  [2023-03-02 11:10:59,408][09136] Avg episode rewards: #0: 4.464, true rewards: #0: 4.064
854
  [2023-03-02 11:10:59,408][09136] Avg episode reward: 4.464, avg true_objective: 4.064
855
  [2023-03-02 11:11:03,282][09136] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
853
  [2023-03-02 11:10:59,408][09136] Avg episode rewards: #0: 4.464, true rewards: #0: 4.064
854
  [2023-03-02 11:10:59,408][09136] Avg episode reward: 4.464, avg true_objective: 4.064
855
  [2023-03-02 11:11:03,282][09136] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!
856
+ [2023-03-02 11:11:22,494][09136] The model has been pushed to https://huggingface.co/nhiro3303/rl_course_vizdoom_health_gathering_supreme
857
+ [2023-03-02 11:41:48,937][09917] Saving configuration to /home/gpu/train_dir/default_experiment/config.json...
858
+ [2023-03-02 11:41:48,938][09917] Rollout worker 0 uses device cpu
859
+ [2023-03-02 11:41:48,938][09917] Rollout worker 1 uses device cpu
860
+ [2023-03-02 11:41:48,938][09917] Rollout worker 2 uses device cpu
861
+ [2023-03-02 11:41:48,938][09917] Rollout worker 3 uses device cpu
862
+ [2023-03-02 11:41:48,938][09917] Rollout worker 4 uses device cpu
863
+ [2023-03-02 11:41:48,938][09917] Rollout worker 5 uses device cpu
864
+ [2023-03-02 11:41:48,938][09917] Rollout worker 6 uses device cpu
865
+ [2023-03-02 11:41:48,938][09917] Rollout worker 7 uses device cpu
866
+ [2023-03-02 11:41:48,966][09917] Using GPUs [0] for process 0 (actually maps to GPUs [0])
867
+ [2023-03-02 11:41:48,966][09917] InferenceWorker_p0-w0: min num requests: 2
868
+ [2023-03-02 11:41:48,983][09917] Starting all processes...
869
+ [2023-03-02 11:41:48,983][09917] Starting process learner_proc0
870
+ [2023-03-02 11:41:49,683][09917] Starting all processes...
871
+ [2023-03-02 11:41:49,686][09975] Using GPUs [0] for process 0 (actually maps to GPUs [0])
872
+ [2023-03-02 11:41:49,686][09975] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for learning process 0
873
+ [2023-03-02 11:41:49,686][09917] Starting process inference_proc0-0
874
+ [2023-03-02 11:41:49,686][09917] Starting process rollout_proc0
875
+ [2023-03-02 11:41:49,686][09917] Starting process rollout_proc1
876
+ [2023-03-02 11:41:49,688][09917] Starting process rollout_proc2
877
+ [2023-03-02 11:41:49,692][09917] Starting process rollout_proc3
878
+ [2023-03-02 11:41:49,694][09917] Starting process rollout_proc4
879
+ [2023-03-02 11:41:49,694][09917] Starting process rollout_proc5
880
+ [2023-03-02 11:41:49,694][09917] Starting process rollout_proc6
881
+ [2023-03-02 11:41:49,694][09917] Starting process rollout_proc7
882
+ [2023-03-02 11:41:49,721][09975] Num visible devices: 1
883
+ [2023-03-02 11:41:49,748][09975] Starting seed is not provided
884
+ [2023-03-02 11:41:49,748][09975] Using GPUs [0] for process 0 (actually maps to GPUs [0])
885
+ [2023-03-02 11:41:49,748][09975] Initializing actor-critic model on device cuda:0
886
+ [2023-03-02 11:41:49,749][09975] RunningMeanStd input shape: (3, 72, 128)
887
+ [2023-03-02 11:41:49,750][09975] RunningMeanStd input shape: (1,)
888
+ [2023-03-02 11:41:49,762][09975] ConvEncoder: input_channels=3
889
+ [2023-03-02 11:41:50,005][09975] Conv encoder output size: 512
890
+ [2023-03-02 11:41:50,005][09975] Policy head output size: 512
891
+ [2023-03-02 11:41:50,018][09975] Created Actor Critic model with architecture:
892
+ [2023-03-02 11:41:50,019][09975] ActorCriticSharedWeights(
893
+ (obs_normalizer): ObservationNormalizer(
894
+ (running_mean_std): RunningMeanStdDictInPlace(
895
+ (running_mean_std): ModuleDict(
896
+ (obs): RunningMeanStdInPlace()
897
+ )
898
+ )
899
+ )
900
+ (returns_normalizer): RecursiveScriptModule(original_name=RunningMeanStdInPlace)
901
+ (encoder): VizdoomEncoder(
902
+ (basic_encoder): ConvEncoder(
903
+ (enc): RecursiveScriptModule(
904
+ original_name=ConvEncoderImpl
905
+ (conv_head): RecursiveScriptModule(
906
+ original_name=Sequential
907
+ (0): RecursiveScriptModule(original_name=Conv2d)
908
+ (1): RecursiveScriptModule(original_name=ELU)
909
+ (2): RecursiveScriptModule(original_name=Conv2d)
910
+ (3): RecursiveScriptModule(original_name=ELU)
911
+ (4): RecursiveScriptModule(original_name=Conv2d)
912
+ (5): RecursiveScriptModule(original_name=ELU)
913
+ )
914
+ (mlp_layers): RecursiveScriptModule(
915
+ original_name=Sequential
916
+ (0): RecursiveScriptModule(original_name=Linear)
917
+ (1): RecursiveScriptModule(original_name=ELU)
918
+ )
919
+ )
920
+ )
921
+ )
922
+ (core): ModelCoreRNN(
923
+ (core): GRU(512, 512)
924
+ )
925
+ (decoder): MlpDecoder(
926
+ (mlp): Identity()
927
+ )
928
+ (critic_linear): Linear(in_features=512, out_features=1, bias=True)
929
+ (action_parameterization): ActionParameterizationDefault(
930
+ (distribution_linear): Linear(in_features=512, out_features=5, bias=True)
931
+ )
932
+ )
933
+ [2023-03-02 11:41:50,847][10004] Worker 0 uses CPU cores [0, 1]
934
+ [2023-03-02 11:41:50,857][10011] Worker 7 uses CPU cores [14, 15]
935
+ [2023-03-02 11:41:50,857][10005] Worker 1 uses CPU cores [2, 3]
936
+ [2023-03-02 11:41:50,857][10009] Worker 5 uses CPU cores [10, 11]
937
+ [2023-03-02 11:41:50,888][10006] Worker 2 uses CPU cores [4, 5]
938
+ [2023-03-02 11:41:50,890][10007] Worker 3 uses CPU cores [6, 7]
939
+ [2023-03-02 11:41:50,891][10010] Worker 6 uses CPU cores [12, 13]
940
+ [2023-03-02 11:41:50,908][10003] Using GPUs [0] for process 0 (actually maps to GPUs [0])
941
+ [2023-03-02 11:41:50,908][10003] Set environment var CUDA_VISIBLE_DEVICES to '0' (GPU indices [0]) for inference process 0
942
+ [2023-03-02 11:41:50,921][10003] Num visible devices: 1
943
+ [2023-03-02 11:41:50,924][10008] Worker 4 uses CPU cores [8, 9]
944
+ [2023-03-02 11:41:52,109][09975] Using optimizer <class 'torch.optim.adam.Adam'>
945
+ [2023-03-02 11:41:52,109][09975] Loading state from checkpoint /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000856_3506176.pth...
946
+ [2023-03-02 11:41:52,129][09975] Loading model from checkpoint
947
+ [2023-03-02 11:41:52,136][09975] Loaded experiment state at self.train_step=856, self.env_steps=3506176
948
+ [2023-03-02 11:41:52,137][09975] Initialized policy 0 weights for model version 856
949
+ [2023-03-02 11:41:52,140][09975] LearnerWorker_p0 finished initialization!
950
+ [2023-03-02 11:41:52,140][09975] Using GPUs [0] for process 0 (actually maps to GPUs [0])
951
+ [2023-03-02 11:41:52,242][10003] Unhandled exception CUDA error: invalid resource handle in evt loop inference_proc0-0_evt_loop
952
+ [2023-03-02 11:41:52,400][09917] Fps is (10 sec: nan, 60 sec: nan, 300 sec: nan). Total num frames: 3506176. Throughput: 0: nan. Samples: 0. Policy #0 lag: (min: -1.0, avg: -1.0, max: -1.0)
953
+ [2023-03-02 11:41:57,400][09917] Fps is (10 sec: 0.0, 60 sec: 0.0, 300 sec: 0.0). Total num frames: 3506176. Throughput: 0: 0.0. Samples: 0. Policy #0 lag: (min: -1.0, avg: -1.0, max: -1.0)
954
+ [2023-03-02 11:42:02,400][09917] Fps is (10 sec: 0.0, 60 sec: 0.0, 300 sec: 0.0). Total num frames: 3506176. Throughput: 0: 0.0. Samples: 0. Policy #0 lag: (min: -1.0, avg: -1.0, max: -1.0)
955
+ [2023-03-02 11:42:07,400][09917] Fps is (10 sec: 0.0, 60 sec: 0.0, 300 sec: 0.0). Total num frames: 3506176. Throughput: 0: 0.0. Samples: 0. Policy #0 lag: (min: -1.0, avg: -1.0, max: -1.0)
956
+ [2023-03-02 11:42:08,483][09917] Keyboard interrupt detected in the event loop EvtLoop [Runner_EvtLoop, process=main process 9917], exiting...
957
+ [2023-03-02 11:42:08,484][10009] Stopping RolloutWorker_w5...
958
+ [2023-03-02 11:42:08,484][10008] Stopping RolloutWorker_w4...
959
+ [2023-03-02 11:42:08,484][09917] Runner profile tree view:
960
+ main_loop: 19.5008
961
+ [2023-03-02 11:42:08,484][09917] Collected {0: 3506176}, FPS: 0.0
962
+ [2023-03-02 11:42:08,484][10005] Stopping RolloutWorker_w1...
963
+ [2023-03-02 11:42:08,484][10006] Stopping RolloutWorker_w2...
964
+ [2023-03-02 11:42:08,484][10010] Stopping RolloutWorker_w6...
965
+ [2023-03-02 11:42:08,484][10009] Loop rollout_proc5_evt_loop terminating...
966
+ [2023-03-02 11:42:08,484][10007] Stopping RolloutWorker_w3...
967
+ [2023-03-02 11:42:08,484][09975] Stopping Batcher_0...
968
+ [2023-03-02 11:42:08,484][10004] Stopping RolloutWorker_w0...
969
+ [2023-03-02 11:42:08,484][10008] Loop rollout_proc4_evt_loop terminating...
970
+ [2023-03-02 11:42:08,484][10011] Stopping RolloutWorker_w7...
971
+ [2023-03-02 11:42:08,484][10006] Loop rollout_proc2_evt_loop terminating...
972
+ [2023-03-02 11:42:08,484][10005] Loop rollout_proc1_evt_loop terminating...
973
+ [2023-03-02 11:42:08,484][10010] Loop rollout_proc6_evt_loop terminating...
974
+ [2023-03-02 11:42:08,484][10007] Loop rollout_proc3_evt_loop terminating...
975
+ [2023-03-02 11:42:08,484][10004] Loop rollout_proc0_evt_loop terminating...
976
+ [2023-03-02 11:42:08,484][10011] Loop rollout_proc7_evt_loop terminating...
977
+ [2023-03-02 11:42:08,484][09975] Loop batcher_evt_loop terminating...
978
+ [2023-03-02 11:42:08,485][09975] Saving /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000856_3506176.pth...
979
+ [2023-03-02 11:42:08,546][09975] Stopping LearnerWorker_p0...
980
+ [2023-03-02 11:42:08,547][09975] Loop learner_proc0_evt_loop terminating...
981
+ [2023-03-02 11:42:08,580][09917] Loading existing experiment configuration from /home/gpu/train_dir/default_experiment/config.json
982
+ [2023-03-02 11:42:08,580][09917] Overriding arg 'num_workers' with value 1 passed from command line
983
+ [2023-03-02 11:42:08,580][09917] Adding new argument 'no_render'=True that is not in the saved config file!
984
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'save_video'=True that is not in the saved config file!
985
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
986
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'video_name'=None that is not in the saved config file!
987
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'max_num_frames'=1000000000.0 that is not in the saved config file!
988
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
989
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'push_to_hub'=False that is not in the saved config file!
990
+ [2023-03-02 11:42:08,581][09917] Adding new argument 'hf_repository'=None that is not in the saved config file!
991
+ [2023-03-02 11:42:08,582][09917] Adding new argument 'policy_index'=0 that is not in the saved config file!
992
+ [2023-03-02 11:42:08,582][09917] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
993
+ [2023-03-02 11:42:08,582][09917] Adding new argument 'train_script'=None that is not in the saved config file!
994
+ [2023-03-02 11:42:08,582][09917] Adding new argument 'enjoy_script'=None that is not in the saved config file!
995
+ [2023-03-02 11:42:08,582][09917] Using frameskip 1 and render_action_repeat=4 for evaluation
996
+ [2023-03-02 11:42:08,592][09917] Doom resolution: 160x120, resize resolution: (128, 72)
997
+ [2023-03-02 11:42:08,593][09917] RunningMeanStd input shape: (3, 72, 128)
998
+ [2023-03-02 11:42:08,593][09917] RunningMeanStd input shape: (1,)
999
+ [2023-03-02 11:42:08,606][09917] ConvEncoder: input_channels=3
1000
+ [2023-03-02 11:42:08,837][09917] Conv encoder output size: 512
1001
+ [2023-03-02 11:42:08,838][09917] Policy head output size: 512
1002
+ [2023-03-02 11:42:10,551][09917] Loading state from checkpoint /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000856_3506176.pth...
1003
+ [2023-03-02 11:42:11,320][09917] Num frames 100...
1004
+ [2023-03-02 11:42:11,449][09917] Num frames 200...
1005
+ [2023-03-02 11:42:11,577][09917] Num frames 300...
1006
+ [2023-03-02 11:42:11,704][09917] Num frames 400...
1007
+ [2023-03-02 11:42:11,818][09917] Avg episode rewards: #0: 5.480, true rewards: #0: 4.480
1008
+ [2023-03-02 11:42:11,818][09917] Avg episode reward: 5.480, avg true_objective: 4.480
1009
+ [2023-03-02 11:42:11,886][09917] Num frames 500...
1010
+ [2023-03-02 11:42:12,014][09917] Num frames 600...
1011
+ [2023-03-02 11:42:12,141][09917] Num frames 700...
1012
+ [2023-03-02 11:42:12,267][09917] Num frames 800...
1013
+ [2023-03-02 11:42:12,404][09917] Avg episode rewards: #0: 5.320, true rewards: #0: 4.320
1014
+ [2023-03-02 11:42:12,404][09917] Avg episode reward: 5.320, avg true_objective: 4.320
1015
+ [2023-03-02 11:42:12,450][09917] Num frames 900...
1016
+ [2023-03-02 11:42:12,576][09917] Num frames 1000...
1017
+ [2023-03-02 11:42:12,701][09917] Num frames 1100...
1018
+ [2023-03-02 11:42:12,827][09917] Num frames 1200...
1019
+ [2023-03-02 11:42:12,940][09917] Avg episode rewards: #0: 4.827, true rewards: #0: 4.160
1020
+ [2023-03-02 11:42:12,940][09917] Avg episode reward: 4.827, avg true_objective: 4.160
1021
+ [2023-03-02 11:42:13,007][09917] Num frames 1300...
1022
+ [2023-03-02 11:42:13,136][09917] Num frames 1400...
1023
+ [2023-03-02 11:42:13,262][09917] Num frames 1500...
1024
+ [2023-03-02 11:42:13,393][09917] Num frames 1600...
1025
+ [2023-03-02 11:42:13,487][09917] Avg episode rewards: #0: 4.580, true rewards: #0: 4.080
1026
+ [2023-03-02 11:42:13,487][09917] Avg episode reward: 4.580, avg true_objective: 4.080
1027
+ [2023-03-02 11:42:13,578][09917] Num frames 1700...
1028
+ [2023-03-02 11:42:13,705][09917] Num frames 1800...
1029
+ [2023-03-02 11:42:13,834][09917] Num frames 1900...
1030
+ [2023-03-02 11:42:13,961][09917] Num frames 2000...
1031
+ [2023-03-02 11:42:14,036][09917] Avg episode rewards: #0: 4.432, true rewards: #0: 4.032
1032
+ [2023-03-02 11:42:14,036][09917] Avg episode reward: 4.432, avg true_objective: 4.032
1033
+ [2023-03-02 11:42:14,156][09917] Num frames 2100...
1034
+ [2023-03-02 11:42:14,296][09917] Num frames 2200...
1035
+ [2023-03-02 11:42:14,484][09917] Num frames 2300...
1036
+ [2023-03-02 11:42:14,624][09917] Num frames 2400...
1037
+ [2023-03-02 11:42:14,675][09917] Avg episode rewards: #0: 4.333, true rewards: #0: 4.000
1038
+ [2023-03-02 11:42:14,675][09917] Avg episode reward: 4.333, avg true_objective: 4.000
1039
+ [2023-03-02 11:42:14,803][09917] Num frames 2500...
1040
+ [2023-03-02 11:42:14,929][09917] Num frames 2600...
1041
+ [2023-03-02 11:42:15,056][09917] Num frames 2700...
1042
+ [2023-03-02 11:42:15,184][09917] Num frames 2800...
1043
+ [2023-03-02 11:42:15,299][09917] Avg episode rewards: #0: 4.497, true rewards: #0: 4.069
1044
+ [2023-03-02 11:42:15,300][09917] Avg episode reward: 4.497, avg true_objective: 4.069
1045
+ [2023-03-02 11:42:15,376][09917] Num frames 2900...
1046
+ [2023-03-02 11:42:15,507][09917] Num frames 3000...
1047
+ [2023-03-02 11:42:15,632][09917] Num frames 3100...
1048
+ [2023-03-02 11:42:15,772][09917] Num frames 3200...
1049
+ [2023-03-02 11:42:15,974][09917] Avg episode rewards: #0: 4.620, true rewards: #0: 4.120
1050
+ [2023-03-02 11:42:15,974][09917] Avg episode reward: 4.620, avg true_objective: 4.120
1051
+ [2023-03-02 11:42:15,982][09917] Num frames 3300...
1052
+ [2023-03-02 11:42:16,136][09917] Num frames 3400...
1053
+ [2023-03-02 11:42:16,291][09917] Num frames 3500...
1054
+ [2023-03-02 11:42:16,451][09917] Num frames 3600...
1055
+ [2023-03-02 11:42:16,637][09917] Avg episode rewards: #0: 4.533, true rewards: #0: 4.089
1056
+ [2023-03-02 11:42:16,638][09917] Avg episode reward: 4.533, avg true_objective: 4.089
1057
+ [2023-03-02 11:42:16,676][09917] Num frames 3700...
1058
+ [2023-03-02 11:42:16,842][09917] Num frames 3800...
1059
+ [2023-03-02 11:42:17,012][09917] Num frames 3900...
1060
+ [2023-03-02 11:42:17,185][09917] Num frames 4000...
1061
+ [2023-03-02 11:42:17,362][09917] Avg episode rewards: #0: 4.464, true rewards: #0: 4.064
1062
+ [2023-03-02 11:42:17,362][09917] Avg episode reward: 4.464, avg true_objective: 4.064
1063
+ [2023-03-02 11:42:21,426][09917] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!
1064
+ [2023-03-02 11:42:32,345][09917] Loading existing experiment configuration from /home/gpu/train_dir/default_experiment/config.json
1065
+ [2023-03-02 11:42:32,345][09917] Overriding arg 'num_workers' with value 1 passed from command line
1066
+ [2023-03-02 11:42:32,345][09917] Adding new argument 'no_render'=True that is not in the saved config file!
1067
+ [2023-03-02 11:42:32,345][09917] Adding new argument 'save_video'=True that is not in the saved config file!
1068
+ [2023-03-02 11:42:32,345][09917] Adding new argument 'video_frames'=1000000000.0 that is not in the saved config file!
1069
+ [2023-03-02 11:42:32,345][09917] Adding new argument 'video_name'=None that is not in the saved config file!
1070
+ [2023-03-02 11:42:32,345][09917] Adding new argument 'max_num_frames'=100000 that is not in the saved config file!
1071
+ [2023-03-02 11:42:32,345][09917] Adding new argument 'max_num_episodes'=10 that is not in the saved config file!
1072
+ [2023-03-02 11:42:32,346][09917] Adding new argument 'push_to_hub'=True that is not in the saved config file!
1073
+ [2023-03-02 11:42:32,346][09917] Adding new argument 'hf_repository'='nhiro3303/rl_course_vizdoom_health_gathering_supreme' that is not in the saved config file!
1074
+ [2023-03-02 11:42:32,346][09917] Adding new argument 'policy_index'=0 that is not in the saved config file!
1075
+ [2023-03-02 11:42:32,346][09917] Adding new argument 'eval_deterministic'=False that is not in the saved config file!
1076
+ [2023-03-02 11:42:32,346][09917] Adding new argument 'train_script'=None that is not in the saved config file!
1077
+ [2023-03-02 11:42:32,346][09917] Adding new argument 'enjoy_script'=None that is not in the saved config file!
1078
+ [2023-03-02 11:42:32,346][09917] Using frameskip 1 and render_action_repeat=4 for evaluation
1079
+ [2023-03-02 11:42:32,349][09917] RunningMeanStd input shape: (3, 72, 128)
1080
+ [2023-03-02 11:42:32,349][09917] RunningMeanStd input shape: (1,)
1081
+ [2023-03-02 11:42:32,356][09917] ConvEncoder: input_channels=3
1082
+ [2023-03-02 11:42:32,379][09917] Conv encoder output size: 512
1083
+ [2023-03-02 11:42:32,379][09917] Policy head output size: 512
1084
+ [2023-03-02 11:42:32,399][09917] Loading state from checkpoint /home/gpu/train_dir/default_experiment/checkpoint_p0/checkpoint_000000856_3506176.pth...
1085
+ [2023-03-02 11:42:32,789][09917] Num frames 100...
1086
+ [2023-03-02 11:42:33,008][09917] Num frames 200...
1087
+ [2023-03-02 11:42:33,216][09917] Num frames 300...
1088
+ [2023-03-02 11:42:33,431][09917] Num frames 400...
1089
+ [2023-03-02 11:42:33,520][09917] Avg episode rewards: #0: 4.160, true rewards: #0: 4.160
1090
+ [2023-03-02 11:42:33,521][09917] Avg episode reward: 4.160, avg true_objective: 4.160
1091
+ [2023-03-02 11:42:33,708][09917] Num frames 500...
1092
+ [2023-03-02 11:42:33,921][09917] Num frames 600...
1093
+ [2023-03-02 11:42:34,148][09917] Num frames 700...
1094
+ [2023-03-02 11:42:34,362][09917] Num frames 800...
1095
+ [2023-03-02 11:42:34,484][09917] Avg episode rewards: #0: 4.660, true rewards: #0: 4.160
1096
+ [2023-03-02 11:42:34,484][09917] Avg episode reward: 4.660, avg true_objective: 4.160
1097
+ [2023-03-02 11:42:34,633][09917] Num frames 900...
1098
+ [2023-03-02 11:42:34,835][09917] Num frames 1000...
1099
+ [2023-03-02 11:42:35,067][09917] Avg episode rewards: #0: 3.960, true rewards: #0: 3.627
1100
+ [2023-03-02 11:42:35,067][09917] Avg episode reward: 3.960, avg true_objective: 3.627
1101
+ [2023-03-02 11:42:35,089][09917] Num frames 1100...
1102
+ [2023-03-02 11:42:35,305][09917] Num frames 1200...
1103
+ [2023-03-02 11:42:35,521][09917] Num frames 1300...
1104
+ [2023-03-02 11:42:35,726][09917] Num frames 1400...
1105
+ [2023-03-02 11:42:35,864][09917] Avg episode rewards: #0: 4.100, true rewards: #0: 3.600
1106
+ [2023-03-02 11:42:35,864][09917] Avg episode reward: 4.100, avg true_objective: 3.600
1107
+ [2023-03-02 11:42:35,991][09917] Num frames 1500...
1108
+ [2023-03-02 11:42:36,195][09917] Num frames 1600...
1109
+ [2023-03-02 11:42:36,414][09917] Num frames 1700...
1110
+ [2023-03-02 11:42:36,616][09917] Num frames 1800...
1111
+ [2023-03-02 11:42:36,718][09917] Avg episode rewards: #0: 4.048, true rewards: #0: 3.648
1112
+ [2023-03-02 11:42:36,718][09917] Avg episode reward: 4.048, avg true_objective: 3.648
1113
+ [2023-03-02 11:42:36,882][09917] Num frames 1900...
1114
+ [2023-03-02 11:42:37,083][09917] Num frames 2000...
1115
+ [2023-03-02 11:42:37,277][09917] Num frames 2100...
1116
+ [2023-03-02 11:42:37,490][09917] Num frames 2200...
1117
+ [2023-03-02 11:42:37,561][09917] Avg episode rewards: #0: 4.013, true rewards: #0: 3.680
1118
+ [2023-03-02 11:42:37,561][09917] Avg episode reward: 4.013, avg true_objective: 3.680
1119
+ [2023-03-02 11:42:37,746][09917] Num frames 2300...
1120
+ [2023-03-02 11:42:38,005][09917] Num frames 2400...
1121
+ [2023-03-02 11:42:38,214][09917] Num frames 2500...
1122
+ [2023-03-02 11:42:38,449][09917] Avg episode rewards: #0: 3.989, true rewards: #0: 3.703
1123
+ [2023-03-02 11:42:38,449][09917] Avg episode reward: 3.989, avg true_objective: 3.703
1124
+ [2023-03-02 11:42:38,466][09917] Num frames 2600...
1125
+ [2023-03-02 11:42:38,670][09917] Num frames 2700...
1126
+ [2023-03-02 11:42:38,865][09917] Num frames 2800...
1127
+ [2023-03-02 11:42:39,070][09917] Num frames 2900...
1128
+ [2023-03-02 11:42:39,290][09917] Avg episode rewards: #0: 3.970, true rewards: #0: 3.720
1129
+ [2023-03-02 11:42:39,290][09917] Avg episode reward: 3.970, avg true_objective: 3.720
1130
+ [2023-03-02 11:42:39,353][09917] Num frames 3000...
1131
+ [2023-03-02 11:42:39,552][09917] Num frames 3100...
1132
+ [2023-03-02 11:42:39,763][09917] Num frames 3200...
1133
+ [2023-03-02 11:42:39,965][09917] Num frames 3300...
1134
+ [2023-03-02 11:42:40,144][09917] Avg episode rewards: #0: 3.956, true rewards: #0: 3.733
1135
+ [2023-03-02 11:42:40,145][09917] Avg episode reward: 3.956, avg true_objective: 3.733
1136
+ [2023-03-02 11:42:40,232][09917] Num frames 3400...
1137
+ [2023-03-02 11:42:40,452][09917] Num frames 3500...
1138
+ [2023-03-02 11:42:40,660][09917] Num frames 3600...
1139
+ [2023-03-02 11:42:40,745][09917] Avg episode rewards: #0: 3.816, true rewards: #0: 3.616
1140
+ [2023-03-02 11:42:40,745][09917] Avg episode reward: 3.816, avg true_objective: 3.616
1141
+ [2023-03-02 11:42:44,193][09917] Replay video saved to /home/gpu/train_dir/default_experiment/replay.mp4!