Réponse apportée
Question about using the nonlinear model predictive control toolbox
nlmpcmove function uses fmincon from Optimization Toolbox. Pleaase take a look at this doc page for more details

environ 3 ans il y a | 0

Réponse apportée
Adaptive MPC for LPV systems
This is possible with Model Predictive Control Toolbox yes, please take a look at this example.

environ 3 ans il y a | 0

Réponse apportée
Is it possible to pass simulink NMPC parameters as struct?
Please take a look at my answer here

environ 3 ans il y a | 0

| A accepté

Réponse apportée
Get nonlinear model from Simulink to Matlab
Hi Daniel, Unfortunately Model Predictive Control Toolbox does not currently support using a Simulink model directly as a predi...

environ 3 ans il y a | 0

| A accepté

Réponse apportée
Binary Manipulated Variable in NLMPC
Unfortunately Model Predictive Control Toolbox does not currently support mixed-integer problems for NLMPC.

environ 3 ans il y a | 0

Réponse apportée
Import Timeseries on Simulink as raw structure and not as single data samples at certain simulation time
It seems to me that you want to implement the MPC preview capability in Simulink. This example actually has a custom previewer b...

environ 3 ans il y a | 0

Réponse apportée
NLMPC state function parameters
Hi, Given that it is not possible to package all variables into a struct and feed them into the MPC controller that way (it's n...

environ 3 ans il y a | 0

| A accepté

Réponse apportée
Can I Generate C Code for TD3 or DDPG agent?
Yes. Here is an example

environ 3 ans il y a | 0

Réponse apportée
Applying reinforcement learning with two continuous actions. During training one varies but the other is virtually static.
You should be able to use a single agent for this task. Since you are using DDPG, the first thing I would check is whether the n...

environ 3 ans il y a | 0

Réponse apportée
How the generated parameters of agent are related PI/PID controller in reinforcement learning algorithm
This example uses an actor network that works as a PI controller. This is possible because we are using a customer layer called ...

environ 3 ans il y a | 0

| A accepté

Réponse apportée
Constantly high negative reward in RL agent
You need to see what the actions generated by the RL Agent block are and how they affect the quadrotor dynamics. That's what it ...

environ 3 ans il y a | 0

| A accepté

Réponse apportée
Load PPO agent and check layer
The neural network architectures are not immediately available from the agent object. You can extract and inspect layers using g...

environ 3 ans il y a | 0

Réponse apportée
cumulative reward in RL AGENT block.
It's not very clear what you are asking, but the cumulative reward changes during each episode. It also depends on the sample ti...

environ 3 ans il y a | 0

Réponse apportée
Sending episode number to simulink during training
The training stats are not available until after training is done, so you cannot use trainingStats.EpisodeIndex during training....

environ 3 ans il y a | 0

| A accepté

Réponse apportée
How to modify actions in experiences during a reinforcement learning training
If you are working in Simulink, you can use the "Last Action" port in the RL Agent block to indicate what was the action that wa...

plus de 3 ans il y a | 1

Réponse apportée
How to get the actor network of a trained policy gradient agent?
Hello, To get the neural network model you can use net = getModel(getActor(agent)) To get learnable parameters you can use g...

plus de 3 ans il y a | 0

| A accepté

Réponse apportée
How to see actions when using the train() function in RL tool box.
Hello, To log action data throughout an episode, you would need to do so from inside the step function of your environment. You...

presque 4 ans il y a | 1

| A accepté

Réponse apportée
Training Quadrotor using PPO agent
Hello, There are multiple things not set up properly, including: 1) The isdone flag seems to be 1 all the time leading to epis...

presque 4 ans il y a | 0

Réponse apportée
How to train RL-DQN agent with varying environment?
What you are describing is actually pretty standard process to create robust policies. To change the driving profiles, you can u...

plus de 4 ans il y a | 2

| A accepté

Réponse apportée
Editing the Q-table before Training in Basic Grid World?
Hello, Please take a look at this link that mentions how you can initialize the table.

plus de 4 ans il y a | 0

| A accepté

Réponse apportée
Could I learn from past data INCLUDING actions? Could I make vector with actions to be used in a certain order?
Hello, If the historical observations do not depend on the actions taken, (think of stock values, or historical power demand), ...

plus de 4 ans il y a | 1

| A accepté

Réponse apportée
update reinforcement policy.m weights
Hello, When you want to perform inference on an RL policy, there is no need to consider rewards. The trained policy already kno...

plus de 4 ans il y a | 0

| A accepté

Réponse apportée
I believe the RL environment template creator has an error in the reset function but I'm not sure
Hello, You are correct the order is wrong. That being said, the order of states depends on your dynamics and how you set up the...

plus de 4 ans il y a | 0

| A accepté

Réponse apportée
What exactly is Episode Q0? What information is it giving?
Q0 is calculated by performing inference on the critic at the beginning of each episode. Effectively, it is a metric that tells ...

plus de 4 ans il y a | 2

| A accepté

Réponse apportée
Resume training of a DQN agent. How to avoid Epsilon from being reset to max value?
Hello, This is currently not possible, but it is a great enhancement idea. I have informed the developers about your request an...

plus de 4 ans il y a | 0

| A accepté

Réponse apportée
Reinforcement learning with Simulink and Simscape
Even outside the thermal domain, you most likely need to start with a simulation model. RL does not need to build that model nec...

plus de 4 ans il y a | 0

Réponse apportée
RL training result very different from the result of 'sim'
Please see this post that explains why simulation results may differ during training and after training. If the simulation resu...

plus de 4 ans il y a | 0

| A accepté

Réponse apportée
RL in dynamic environment
The following example seems relevant, please take a look: https://www.mathworks.com/help/robotics/ug/avoid-obstacles-using-rein...

plus de 4 ans il y a | 0

| A accepté

Réponse apportée
MPC Controller giving nice performance during designing but fails on testing
Hello, It sounds to me that the issue is with the linearized model. When you are exporting the controller from MPC Designer, yo...

plus de 4 ans il y a | 0

Réponse apportée
What is in a reinforcement learning saved agent .mat file
Why don't you load the file and check? When you saved the agen tin the .mat file, did you save anything else with it? Are you m...

presque 5 ans il y a | 0

Charger plus