rlAdditiveNoisePolicy
Policy object to generate continuous noisy actions for custom training loops
Since R2022a
Description
This object implements an additive noise policy, which returns continuous
deterministic actions with added noise, given an input observation. You can create an
rlAdditiveNoisePolicy
object from an rlContinuousDeterministicActor
or extract it from an rlDDPGAgent
or rlTD3Agent
. You can
then train the policy object using a custom training loop. If
UseNoisyAction
is set to 0
the policy does not
explore. This object is not compatible with generatePolicyBlock
and generatePolicyFunction
. For more information on policies and value functions,
see Create Policies and Value Functions.
Creation
Syntax
Description
creates the additive noise policy object policy
= rlAdditiveNoisePolicy(actor
)policy
from the continuous
deterministic actor actor
. It also sets the
Actor
property of policy
to the input
argument actor
.
specifies the type of noise distribution for the policy. policy
= rlAdditiveNoisePolicy(actor
,NoiseType=noiseType
)noiseType
can be either "gaussian"
(Gaussian noise) or "ou"
(Ornstein-Uhlenbeck noise). This syntax also sets the NoiseType
property of policy
to the input argument
noiseType
.
Properties
Object Functions
getAction | Obtain action from agent, actor, or policy object given environment observations |
getLearnableParameters | Obtain learnable parameter values from agent, function approximator, or policy object |
reset | Reset environment, agent, experience buffer, or policy object |
setLearnableParameters | Set learnable parameter values of agent, function approximator, or policy object |
Examples
Version History
Introduced in R2022a
See Also
Functions
getGreedyPolicy
|getExplorationPolicy
|generatePolicyBlock
|generatePolicyFunction
|getAction
|getLearnableParameters
|setLearnableParameters
Objects
rlMaxQPolicy
|rlEpsilonGreedyPolicy
|rlDeterministicActorPolicy
|rlStochasticActorPolicy
|rlHybridStochasticActorPolicy
|rlTD3Agent
|rlDDPGAgent