已回答
Question about using the nonlinear model predictive control toolbox
nlmpcmove function uses fmincon from Optimization Toolbox. Pleaase take a look at this doc page for more details

2 years 前 | 0

已回答
Adaptive MPC for LPV systems
This is possible with Model Predictive Control Toolbox yes, please take a look at this example.

2 years 前 | 0

已回答
Is it possible to pass simulink NMPC parameters as struct?
Please take a look at my answer here

2 years 前 | 0

| 已接受

已回答
Get nonlinear model from Simulink to Matlab
Hi Daniel, Unfortunately Model Predictive Control Toolbox does not currently support using a Simulink model directly as a predi...

2 years 前 | 0

| 已接受

已回答
Binary Manipulated Variable in NLMPC
Unfortunately Model Predictive Control Toolbox does not currently support mixed-integer problems for NLMPC.

2 years 前 | 0

已回答
Import Timeseries on Simulink as raw structure and not as single data samples at certain simulation time
It seems to me that you want to implement the MPC preview capability in Simulink. This example actually has a custom previewer b...

2 years 前 | 0

已回答
NLMPC state function parameters
Hi, Given that it is not possible to package all variables into a struct and feed them into the MPC controller that way (it's n...

2 years 前 | 0

| 已接受

已回答
Can I Generate C Code for TD3 or DDPG agent?
Yes. Here is an example

2 years 前 | 0

已回答
Applying reinforcement learning with two continuous actions. During training one varies but the other is virtually static.
You should be able to use a single agent for this task. Since you are using DDPG, the first thing I would check is whether the n...

2 years 前 | 0

已回答
How the generated parameters of agent are related PI/PID controller in reinforcement learning algorithm
This example uses an actor network that works as a PI controller. This is possible because we are using a customer layer called ...

2 years 前 | 0

| 已接受

已回答
Constantly high negative reward in RL agent
You need to see what the actions generated by the RL Agent block are and how they affect the quadrotor dynamics. That's what it ...

2 years 前 | 0

| 已接受

已回答
Load PPO agent and check layer
The neural network architectures are not immediately available from the agent object. You can extract and inspect layers using g...

2 years 前 | 0

已回答
cumulative reward in RL AGENT block.
It's not very clear what you are asking, but the cumulative reward changes during each episode. It also depends on the sample ti...

2 years 前 | 0

已回答
Sending episode number to simulink during training
The training stats are not available until after training is done, so you cannot use trainingStats.EpisodeIndex during training....

2 years 前 | 0

| 已接受

已回答
How to modify actions in experiences during a reinforcement learning training
If you are working in Simulink, you can use the "Last Action" port in the RL Agent block to indicate what was the action that wa...

3 years 前 | 1

已回答
How to get the actor network of a trained policy gradient agent?
Hello, To get the neural network model you can use net = getModel(getActor(agent)) To get learnable parameters you can use g...

3 years 前 | 0

| 已接受

已回答
How to see actions when using the train() function in RL tool box.
Hello, To log action data throughout an episode, you would need to do so from inside the step function of your environment. You...

3 years 前 | 1

| 已接受

已回答
Training Quadrotor using PPO agent
Hello, There are multiple things not set up properly, including: 1) The isdone flag seems to be 1 all the time leading to epis...

3 years 前 | 0

已回答
How to train RL-DQN agent with varying environment?
What you are describing is actually pretty standard process to create robust policies. To change the driving profiles, you can u...

4 years 前 | 2

| 已接受

已回答
Editing the Q-table before Training in Basic Grid World?
Hello, Please take a look at this link that mentions how you can initialize the table.

4 years 前 | 0

| 已接受

已回答
Could I learn from past data INCLUDING actions? Could I make vector with actions to be used in a certain order?
Hello, If the historical observations do not depend on the actions taken, (think of stock values, or historical power demand), ...

4 years 前 | 1

| 已接受

已回答
update reinforcement policy.m weights
Hello, When you want to perform inference on an RL policy, there is no need to consider rewards. The trained policy already kno...

4 years 前 | 0

| 已接受

已回答
I believe the RL environment template creator has an error in the reset function but I'm not sure
Hello, You are correct the order is wrong. That being said, the order of states depends on your dynamics and how you set up the...

4 years 前 | 0

| 已接受

已回答
What exactly is Episode Q0? What information is it giving?
Q0 is calculated by performing inference on the critic at the beginning of each episode. Effectively, it is a metric that tells ...

4 years 前 | 2

| 已接受

已回答
Resume training of a DQN agent. How to avoid Epsilon from being reset to max value?
Hello, This is currently not possible, but it is a great enhancement idea. I have informed the developers about your request an...

4 years 前 | 0

| 已接受

已回答
Reinforcement learning with Simulink and Simscape
Even outside the thermal domain, you most likely need to start with a simulation model. RL does not need to build that model nec...

4 years 前 | 0

已回答
RL training result very different from the result of 'sim'
Please see this post that explains why simulation results may differ during training and after training. If the simulation resu...

4 years 前 | 0

| 已接受

已回答
RL in dynamic environment
The following example seems relevant, please take a look: https://www.mathworks.com/help/robotics/ug/avoid-obstacles-using-rein...

4 years 前 | 0

| 已接受

已回答
MPC Controller giving nice performance during designing but fails on testing
Hello, It sounds to me that the issue is with the linearized model. When you are exporting the controller from MPC Designer, yo...

4 years 前 | 0

已回答
What is in a reinforcement learning saved agent .mat file
Why don't you load the file and check? When you saved the agen tin the .mat file, did you save anything else with it? Are you m...

4 years 前 | 0

加载更多