ppo-Pyramid-v1 / run_logs /timers.json
kraken2404's picture
Second Version
5af10f6
{
"name": "root",
"gauges": {
"Pyramids.Policy.Entropy.mean": {
"value": 0.2454661726951599,
"min": 0.2438683658838272,
"max": 1.5889166593551636,
"count": 46
},
"Pyramids.Policy.Entropy.sum": {
"value": 7391.4775390625,
"min": 7195.09228515625,
"max": 48201.375,
"count": 46
},
"Pyramids.Step.mean": {
"value": 1379901.0,
"min": 29910.0,
"max": 1379901.0,
"count": 46
},
"Pyramids.Step.sum": {
"value": 1379901.0,
"min": 29910.0,
"max": 1379901.0,
"count": 46
},
"Pyramids.Policy.ExtrinsicValueEstimate.mean": {
"value": 0.6779295802116394,
"min": -0.09292546659708023,
"max": 0.7612824440002441,
"count": 46
},
"Pyramids.Policy.ExtrinsicValueEstimate.sum": {
"value": 193.20993041992188,
"min": -22.395036697387695,
"max": 227.62344360351562,
"count": 46
},
"Pyramids.Policy.RndValueEstimate.mean": {
"value": 0.06917993724346161,
"min": -0.0011057952651754022,
"max": 0.3723164498806,
"count": 46
},
"Pyramids.Policy.RndValueEstimate.sum": {
"value": 19.71628189086914,
"min": -0.2941415309906006,
"max": 88.23899841308594,
"count": 46
},
"Pyramids.Losses.PolicyLoss.mean": {
"value": 0.0715606110045753,
"min": 0.06373118343569027,
"max": 0.07409594399810193,
"count": 46
},
"Pyramids.Losses.PolicyLoss.sum": {
"value": 1.001848554064054,
"min": 0.4595870365823242,
"max": 1.0637401472389079,
"count": 46
},
"Pyramids.Losses.ValueLoss.mean": {
"value": 0.015125273911204811,
"min": 0.0010844276371333522,
"max": 0.016129469255813295,
"count": 46
},
"Pyramids.Losses.ValueLoss.sum": {
"value": 0.21175383475686735,
"min": 0.010844276371333521,
"max": 0.24194203883719942,
"count": 46
},
"Pyramids.Policy.LearningRate.mean": {
"value": 7.450995475551021e-06,
"min": 7.450995475551021e-06,
"max": 0.00029659622562438776,
"count": 46
},
"Pyramids.Policy.LearningRate.sum": {
"value": 0.0001043139366577143,
"min": 0.0001043139366577143,
"max": 0.0037952302777804277,
"count": 46
},
"Pyramids.Policy.Epsilon.mean": {
"value": 0.1,
"min": 0.09999999999999998,
"max": 0.10000000000000002,
"count": 46
},
"Pyramids.Policy.Epsilon.sum": {
"value": 1.4000000000000001,
"min": 0.7,
"max": 1.5000000000000002,
"count": 46
},
"Pyramids.Policy.Beta.mean": {
"value": 0.0002581149020408164,
"min": 0.0002581149020408164,
"max": 0.009886654275510205,
"count": 46
},
"Pyramids.Policy.Beta.sum": {
"value": 0.0036136086285714296,
"min": 0.0036136086285714296,
"max": 0.12652116375714287,
"count": 46
},
"Pyramids.Losses.RNDLoss.mean": {
"value": 0.010419806465506554,
"min": 0.010419806465506554,
"max": 0.36617204546928406,
"count": 46
},
"Pyramids.Losses.RNDLoss.sum": {
"value": 0.14587728679180145,
"min": 0.14587728679180145,
"max": 2.563204288482666,
"count": 46
},
"Pyramids.Environment.EpisodeLength.mean": {
"value": 262.4824561403509,
"min": 242.04132231404958,
"max": 999.0,
"count": 46
},
"Pyramids.Environment.EpisodeLength.sum": {
"value": 29923.0,
"min": 16581.0,
"max": 33378.0,
"count": 46
},
"Pyramids.Environment.CumulativeReward.mean": {
"value": 1.7024122663495833,
"min": -0.999962551984936,
"max": 1.7566718604648486,
"count": 46
},
"Pyramids.Environment.CumulativeReward.sum": {
"value": 194.0749983638525,
"min": -31.998801663517952,
"max": 224.85399813950062,
"count": 46
},
"Pyramids.Policy.ExtrinsicReward.mean": {
"value": 1.7024122663495833,
"min": -0.999962551984936,
"max": 1.7566718604648486,
"count": 46
},
"Pyramids.Policy.ExtrinsicReward.sum": {
"value": 194.0749983638525,
"min": -31.998801663517952,
"max": 224.85399813950062,
"count": 46
},
"Pyramids.Policy.RndReward.mean": {
"value": 0.028350559796563812,
"min": 0.026500153520542837,
"max": 6.862230844357434,
"count": 46
},
"Pyramids.Policy.RndReward.sum": {
"value": 3.2319638168082747,
"min": 3.1286544502800098,
"max": 116.65792435407639,
"count": 46
},
"Pyramids.IsTraining.mean": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 46
},
"Pyramids.IsTraining.sum": {
"value": 1.0,
"min": 1.0,
"max": 1.0,
"count": 46
}
},
"metadata": {
"timer_format_version": "0.1.0",
"start_time_seconds": "1679229482",
"python_version": "3.9.16 (main, Dec 7 2022, 01:11:51) \n[GCC 9.4.0]",
"command_line_arguments": "/usr/local/bin/mlagents-learn ./config/ppo/PyramidsRND.yaml --env=./training-envs-executables/linux/Pyramids/Pyramids --run-id=PyramidsTraining2 --no-graphics",
"mlagents_version": "0.31.0.dev0",
"mlagents_envs_version": "0.31.0.dev0",
"communication_protocol_version": "1.5.0",
"pytorch_version": "1.11.0+cu102",
"numpy_version": "1.21.2",
"end_time_seconds": "1679232788"
},
"total": 3306.4794070059997,
"count": 1,
"self": 0.8367734960011148,
"children": {
"run_training.setup": {
"total": 0.17392283699882682,
"count": 1,
"self": 0.17392283699882682
},
"TrainerController.start_learning": {
"total": 3305.468710673,
"count": 1,
"self": 2.3896251441310596,
"children": {
"TrainerController._reset_env": {
"total": 7.0244827779988555,
"count": 1,
"self": 7.0244827779988555
},
"TrainerController.advance": {
"total": 3295.9087750568706,
"count": 90145,
"self": 2.4606926954766095,
"children": {
"env_step": {
"total": 2407.1799921052516,
"count": 90145,
"self": 2236.1181065479086,
"children": {
"SubprocessEnvManager._take_step": {
"total": 169.6538362672436,
"count": 90145,
"self": 7.15606690705863,
"children": {
"TorchPolicy.evaluate": {
"total": 162.49776936018498,
"count": 87560,
"self": 162.49776936018498
}
}
},
"workers": {
"total": 1.4080492900993704,
"count": 90145,
"self": 0.0,
"children": {
"worker_root": {
"total": 3297.432015708957,
"count": 90145,
"is_parallel": true,
"self": 1241.269257969976,
"children": {
"run_training.setup": {
"total": 0.0,
"count": 0,
"is_parallel": true,
"self": 0.0,
"children": {
"steps_from_proto": {
"total": 0.002846290999514167,
"count": 1,
"is_parallel": true,
"self": 0.0008743859980313573,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.00197190500148281,
"count": 8,
"is_parallel": true,
"self": 0.00197190500148281
}
}
},
"UnityEnvironment.step": {
"total": 0.05124855100075365,
"count": 1,
"is_parallel": true,
"self": 0.0005672670013154857,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 0.00048581000010017306,
"count": 1,
"is_parallel": true,
"self": 0.00048581000010017306
},
"communicator.exchange": {
"total": 0.048365937000198755,
"count": 1,
"is_parallel": true,
"self": 0.048365937000198755
},
"steps_from_proto": {
"total": 0.0018295369991392363,
"count": 1,
"is_parallel": true,
"self": 0.00040743300087342504,
"children": {
"_process_rank_one_or_two_observation": {
"total": 0.0014221039982658112,
"count": 8,
"is_parallel": true,
"self": 0.0014221039982658112
}
}
}
}
}
}
},
"UnityEnvironment.step": {
"total": 2056.162757738981,
"count": 90144,
"is_parallel": true,
"self": 46.60205633400801,
"children": {
"UnityEnvironment._generate_step_input": {
"total": 33.657413851666206,
"count": 90144,
"is_parallel": true,
"self": 33.657413851666206
},
"communicator.exchange": {
"total": 1838.478166507315,
"count": 90144,
"is_parallel": true,
"self": 1838.478166507315
},
"steps_from_proto": {
"total": 137.42512104599155,
"count": 90144,
"is_parallel": true,
"self": 30.324248896702557,
"children": {
"_process_rank_one_or_two_observation": {
"total": 107.10087214928899,
"count": 721152,
"is_parallel": true,
"self": 107.10087214928899
}
}
}
}
}
}
}
}
}
}
},
"trainer_advance": {
"total": 886.2680902561424,
"count": 90145,
"self": 4.6611480855317495,
"children": {
"process_trajectory": {
"total": 173.06847194963302,
"count": 90145,
"self": 172.8570300576357,
"children": {
"RLTrainer._checkpoint": {
"total": 0.21144189199731045,
"count": 2,
"self": 0.21144189199731045
}
}
},
"_update_policy": {
"total": 708.5384702209776,
"count": 641,
"self": 447.93973490305325,
"children": {
"TorchPPOOptimizer.update": {
"total": 260.59873531792437,
"count": 31875,
"self": 260.59873531792437
}
}
}
}
}
}
},
"trainer_threads": {
"total": 1.4360011846292764e-06,
"count": 1,
"self": 1.4360011846292764e-06
},
"TrainerController._save_models": {
"total": 0.14582625799812376,
"count": 1,
"self": 0.0019542619957064744,
"children": {
"RLTrainer._checkpoint": {
"total": 0.1438719960024173,
"count": 1,
"self": 0.1438719960024173
}
}
}
}
}
}
}