I have some questions about the minibatchsize attribute of PPO+LSTM

9 次查看(过去 30 天)
Hello everyone
I found this sentence when I searched the content of PPO+LSTM in the official MATLAB documentation:“For a PPO agent, the trajectory length is the MiniBatchSize property of its options object” ,But I have some doubts about this statement。When using PPO+LSTM, does the agent no longer need to sample minbatch from the current experience sequence? How to understand "For a PPO agent, the trajectory length is the MiniBatchSize property of its options object".
Hope to get your answer

回答(1 个)

Aneela
Aneela 2024-6-5
Hi xiang,
“MiniBatchSize” refers to the number of samples from experience replay buffer that are used for one iteration of training.
  • For non-recurrent neural networks, these samples are selected randomly because the network treats each input independently.
  • LSTMs require a sequence of experiences to effectively learn temporal features.
  • When using PPO with LSTM, the agent emphasizes on managing sequences of experiences to leverage the LSTM's ability to learn from temporally dependent data.
“For a PPO agent, the trajectory length is the “MiniBatchSize” property of its options object”:
  • A trajectory is a sequence of states, actions, and rewards that an agent experiences in the environment from the start of an episode until a terminal state.
  • The agent learns from trajectories of experiences.
  • The “MiniBatchSize” value specifies the length of these trajectories.
  • If “MiniBatchSize” is set to 50, the LSTM network will be trained on trajectories of experiences where each trajectory is 50 steps long.
For more information on PPO Agent options, please refer to the following MathWorks documentation: https://www.mathworks.com/help/reinforcement-learning/ref/rl.option.rlppoagentoptions.html

类别

Help CenterFile Exchange 中查找有关 Training and Simulation 的更多信息

标签

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by