Study Log (2021.06)
2021-06-01
- S-K RL
- train_FT10_ppo_node_only.py
- do_simulate_on_aggregated_state()
- value_loss, action_loss, dist_entropy = agent.fit(eval=0, reward_setting=’utilization’, device=device, return_scaled=False)
- eval_performance = evaluate_agent_on_aggregated_state(simulator=sim, agent=agent, device=’cpu’, mode=’node_mode’)
- val_performance = validation(agent, path, mode=’node_mode’)
- pyjssp 버전 구분
- GNN-MARL Lastest용
- GNN-MARL Stable용
- train_FT10_ppo_node_only.py
Template
- Fundamental of Reinforcement Learning
- Chapter #.
- 모두를 위한 머신러닝/딥러닝 강의
- Lecture #.
- UCL Course on RL
- Lecture #.
- Reinforcement Learning
- Page #.
- 팡요랩
- 강화학습 1강 - 강화학습 introduction
- 강화학습 2강 - Markov Decision Process
- 강화학습 3강 - Planning by Dynamic Programming
- 강화학습 4강 - Model Free Prediction
- 강화학습 5강 - Model Free Control
- 강화학습 6강 - Value Function Approximation
- 강화학습 7강 - Policy Gradient
- 강화학습 8강 - Integrating Learning and Planning
- 강화학습 9강 - Exploration and Exploitation
- 강화학습 10강 - Classic Games
- Pattern Recognition & Machine Learning
- S-K RL
- multi_step_actor
Comments