rlDeterministicActorPolicy
Policy object to generate continuous deterministic actions for custom training loops and application deployment
Since R2022a
Description
This object implements a deterministic policy, which returns continuous
deterministic actions given an input observation. You can create an
rlDeterministicActorPolicy
object from an rlContinuousDeterministicActor
or extract it from an rlDDPGAgent
or rlTD3Agent
. You can
then train the policy object using a custom training loop or deploy it for your application
using generatePolicyBlock
or generatePolicyFunction
. This policy is always deterministic and does not perform
any exploration. For more information on policies and value functions, see Create Policies and Value Functions.
Creation
Description
creates the deterministic actor policy object policy
= rlDeterministicActorPolicy(actor
)policy
from the
continuous deterministic actor actor
. It also sets the
Actor
property of policy
to the input
argument actor
.
Properties
Object Functions
generatePolicyBlock | Generate Simulink block that evaluates policy of an agent or policy object |
generatePolicyFunction | Generate MATLAB function that evaluates policy of an agent or policy object |
getAction | Obtain action from agent, actor, or policy object given environment observations |
getLearnableParameters | Obtain learnable parameter values from agent, function approximator, or policy object |
reset | Reset environment, agent, experience buffer, or policy object |
setLearnableParameters | Set learnable parameter values of agent, function approximator, or policy object |
Examples
Version History
Introduced in R2022a
See Also
Functions
getGreedyPolicy
|getExplorationPolicy
|generatePolicyBlock
|generatePolicyFunction
|getAction
|getLearnableParameters
|setLearnableParameters
Objects
rlMaxQPolicy
|rlEpsilonGreedyPolicy
|rlAdditiveNoisePolicy
|rlStochasticActorPolicy
|rlHybridStochasticActorPolicy
|rlTD3Agent
|rlDDPGAgent