Hi Andrea,
Converting the GridWorld environment from discrete to continuous actions involves considerable effort and a deep understanding of reinforcement learning (RL) principles.
Here are key aspects to consider:
- Action representation: in the original GridWorld, actions are discrete and represent movements in cardinal directions (eg. North, South, East, West). For continuous actions, you would need to define how actions are represented, such as using vectors to denote directions and magnitude.
- State transitions: You would need to develop a new method to calculate the next step based on the continuous action taken. You will also need to handle collision and boundary conditions as these will get more complicated compared to simple obstacle collisions in case of discrete values.
- The reward structure may also need adjustments. In discrete GridWorld, rewards are typically assigned based on reaching certain cells. With continuous actions, rewards could be based on distances to objectives, with more granular adjustments.
- Most traditional RL algorithms used with GridWorld are designed for discrete action spaces (e.g., Q-Learning). Continuous action spaces often require different algorithms, such as Deep Deterministic Policy Gradient (DDPG), Proximal Policy Optimization (PPO), or Soft Actor-Critic (SAC), which are more complex and involve neural networks to approximate policy and/or value functions.
I hope this helps.