Réponse apportée
Design an actor critic network for non-image inputs
I may be missing something but why don't you frame your observations as a [4 1] vector? That way it would be consistent with how...

environ 2 ans il y a | 0

Réponse apportée
Pausing reinforcement learning by forcing
The proper way to stop it would be through the Episode Manager (top right of the window). Does this not work for you?

environ 2 ans il y a | 0

| A accepté

Réponse apportée
Why DQN training always fails to converge to the optimal value
What I am seeing here is that the average reward tends to converge to the Q0 profile which is the expected behavior of a converg...

environ 2 ans il y a | 0

Réponse apportée
My reinforcement learning simulation runs for only 0 steps and 0 times in Simulink. I am not getting any error messages so I cannot pinpoint the issue, so I decided to ask.
It likely has to do with the priority of execution of the data store blocks. I would look more into it, but honestly I think you...

environ 2 ans il y a | 0

Réponse apportée
PPO and LSTM agent creation
Hi, With lstm policies, BOTH the actor and the critic should have lstm layers. That's why you are getting this error. LSTM po...

environ 2 ans il y a | 0

Réponse apportée
Matlab reinforcement learning stop working after a while
125 episodes are not that many. Is it always freezing on that episode? Given that you are not getting any errors/crashes, my hun...

environ 2 ans il y a | 0

Réponse apportée
DDPG with LSTM layer fails?
Hello, I see a couple of things wrong with the current architecture (could be more): 1) When you use the lstm layer, the input...

environ 2 ans il y a | 0

| A accepté

Réponse apportée
Could you help clarify the terminology and usage of Exploratory Policy and Exploratory Model in TD3 Reinforcement Learning
The answer above is correct with a small caveat: Even if UseExploratorypolicy is set = 0, the agent will still explore during t...

environ 2 ans il y a | 0

| A accepté

Réponse apportée
Load a pretrained neural network object in rlNeuralNetworkEnvironment
Hi Vasu, You can use a pretrained environment model with MBPO agent as follows: 1) Create a rlContinuousDeterministicTransitio...

environ 2 ans il y a | 0

Réponse apportée
Initial condition setting mPC
This should be helpful: https://www.mathworks.com/help/mpc/ref/mpcstate.html

environ 2 ans il y a | 0

Réponse apportée
is it possible to use neural network prediction model in Adaptive MPC ?
You should be able to do that, but we don't have an example to show the workflow unfortunately. You would first need to use aut...

environ 2 ans il y a | 0

| A accepté

Réponse apportée
Make Jacobian as a function in Simulink
I would first use automatic differentiation to get the A,B,C,D matrices using this command. Then you would need to take the gen...

environ 2 ans il y a | 0

Réponse apportée
RL DDPG Actions have high oscillation
Hi, The noise options you are mentioning are only used during training and are essential for exploration. If the plots you are ...

plus de 2 ans il y a | 1

Réponse apportée
I want to know how to understand the sampling time ts of the agent in reinforcement learning, and how its value will affect the training and results?
Think of it as how often you need the agent to make decisions. Sometimes you need to make decisions very frequently (e.g. if you...

plus de 2 ans il y a | 0

| A accepté

Réponse apportée
An Error in LSTM Network Training Using Reinforcement Learning Toolbox
I believe that both the actor and the critic need to be LSTM networks. In your case only the critic is

plus de 2 ans il y a | 0

| A accepté

Réponse apportée
Linear Model Predictive Control with state feedback
The most straightforward way would be to assume that all your states are outputs, so set the C to be the identity matrix.

plus de 2 ans il y a | 0

Réponse apportée
How to soften hard constraints on states in non-linear MPC?
Here is an example that shows how to do that with multi-stage MPC: https://www.mathworks.com/help/mpc/ug/use-multistage-mpc-wit...

plus de 2 ans il y a | 0

Réponse apportée
RL: Continuous action space, but within a desired range(use PPO)
You can always clip the agent output on the environment side. PPO is stochastic so the upper and lower limits are not guaranteed...

plus de 2 ans il y a | 0

Réponse apportée
How to add custom environment for Reinforcement learning toolbox?
Why do you need a 3d world for this problem? Unless you consider the z dimension (e.g. if you do planning for UAVs), you only ne...

plus de 2 ans il y a | 0

Réponse apportée
Reinforcement Learning - PPO agent with hybrid action space
Hello, The workaround you suggested makes sense to me. Unfortunately though, bullet #3 is not currently supported. You cannot d...

plus de 2 ans il y a | 0

Réponse apportée
How to make the agent automatically pause after training to a certain number of turns
You can actually automate this process without having to stop training if you are using Simulink. There is an port 'Use external...

plus de 2 ans il y a | 0

Réponse apportée
Scaling in the QP formulation
I think it's likely a typo. Scaling factors should be included in Wu, Wdu and Wy along with the individual weights.

plus de 2 ans il y a | 0

| A accepté

Réponse apportée
For a Non linear Model predictive control toolbox, how can I give it a mathematical model, which is a subsystem defined in Simulink?
Unfortunately, this is currently not supported but our development team is aware of this request. If you don't have an analytica...

plus de 2 ans il y a | 0

| A accepté

Réponse apportée
Constraints violation in MPC design
The more constraints you are adding, the smaller the solutions space, the harder it gets for the solver to find a solution. If y...

plus de 2 ans il y a | 0

| A accepté

Réponse apportée
Is there any way to graph the MPC cost function?
If you are designing the controller in MATLAB, all 'move' functions have an info argument which containts the cost value (see he...

plus de 2 ans il y a | 0

Réponse apportée
The code for nmpc is not working even if i put it in same folder ?
I wasn't able to reproduce the error. Both state and output functions must be in different .m files. I did that and put everythi...

plus de 2 ans il y a | 1

| A accepté

Réponse apportée
Problem with LSTM and PPO reinforcement learning
Hi, When you set up LSTM networks, the input layer needs to be the sequenceInput layer. See for example here: https://www.math...

plus de 2 ans il y a | 0

Réponse apportée
QP formulation from the MPC toolbox
We are currently using the dense formula as you mentioned, but also working on adding support for sparse problems. The following...

plus de 2 ans il y a | 0

| A accepté

Réponse apportée
Hierarchical Reinforcement Learning in Matlab
Hierarchical RL basically comes down to breaking a compelx problem down into smaller 'easier to train' problems, and potentially...

plus de 2 ans il y a | 0

Réponse apportée
Tune PI Controller Using Reinforcement Learning
I also replied to the other thread. The fullyConnectedPILayer is a custom layer provided in the example - you can open it and se...

plus de 2 ans il y a | 0

| A accepté

Charger plus